split stringclasses 1 value | image_id stringlengths 12 25 | file_name stringlengths 16 29 | image_info dict | caption_info dict | mask_annotations listlengths 2 52 | categories listlengths 1 1 |
|---|---|---|---|---|---|---|
train | 2103_5qvVtB7fqXE_00005163 | 2103_5qvVtB7fqXE_00005163.jpg | {
"data_source": "VIPSeg",
"file_name": "2103_5qvVtB7fqXE_00005163.jpg",
"height": 720,
"id": "2103_5qvVtB7fqXE_00005163",
"width": 1281
} | {
"caption": "A person wearing a striped shirt is reaching towards a black computer monitor that sits on a light brown wooden desk. Next to the monitor is a silver computer tower with various cables plugged into its back, and a black UPS device is positioned to the left of the computer tower. A black keyboard and a black mouse are also on the desk in front of the monitor. The white wall is in the background, with a dark brown door partially visible on the right.",
"caption_ann": "A <3:person wearing a striped shirt> is reaching towards a <5:black computer monitor> that sits on a <4:light brown wooden desk>. Next to the monitor is a <6:silver computer tower> with various cables plugged into its back, and a <1:black UPS device> is positioned to the left of the computer tower. A <7:black keyboard> and a <7:black mouse> are also on the desk in front of the monitor. The <0:white wall> is in the background, with a <2:dark brown door> partially visible on the right.",
"id": 1700,
"image_id": "2103_5qvVtB7fqXE_00005163",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "person wearing a striped shirt"
},
{
"mask_ids": [
5
],
"txt_desc": "black computer monitor"
},
{
"mask_ids": [
4
],
"txt_desc": "light brown wooden desk"
},
{
"mask_ids": [
6
],
"txt_desc": "silver computer tower"
},
{
"mask_ids": [
1
],
"txt_desc": "black UPS device"
},
{
"mask_ids": [
7
],
"txt_desc": "black keyboard"
},
{
"mask_ids": [
7
],
"txt_desc": "black mouse"
},
{
"mask_ids": [
0
],
"txt_desc": "white wall"
},
{
"mask_ids": [
2
],
"txt_desc": "dark brown door"
}
],
"labels": [
"wall",
"other_electronic_product",
"door",
"person",
"table_or_desk",
"screen_or_television",
"computer",
"keyboard"
]
} | [
{
"area": 251819,
"bbox": [
0,
0,
1272,
720
],
"category_id": 0,
"id": 18432,
"image_id": "2103_5qvVtB7fqXE_00005163",
"iscrowd": 0,
"segmentation": {
"counts": "0e>n6m0000O1000000N20000O100000000N20000O1000000N20000O100000000N20000O10000N2000000O10000N2... | [
{
"id": 1,
"name": "object"
}
] |
train | 2112_N3YcO7YTOBE_00000483 | 2112_N3YcO7YTOBE_00000483.jpg | {
"data_source": "VIPSeg",
"file_name": "2112_N3YcO7YTOBE_00000483.jpg",
"height": 720,
"id": "2112_N3YcO7YTOBE_00000483",
"width": 1280
} | {
"caption": "A man in dark shorts and a light gray full-sleeved pullover is performing a handstand on the steps of an upward-moving escalator, framed by the metal and glass railings, under a slatted, lighted ceiling and a long fluorescent lamp and two identical round shaped lamps, near a directory sign reading \"2nd FLOOR\" and \"3rd FLOOR\" against a reddish-brown wall, with a partially visible person visible behind the man.",
"caption_ann": "A <6:man in dark shorts and a light gray full-sleeved pullover> is performing a handstand on the steps of an <2:upward-moving escalator>, framed by the <3:metal and glass railings>, under a <1:slatted, lighted ceiling> and a <5:long fluorescent lamp and two identical round shaped lamps>, near a <4:directory sign reading \"2nd FLOOR\" and \"3rd FLOOR\"> against a <0:reddish-brown wall>, with a <7:partially visible person> visible behind the <6:man>.",
"id": 1701,
"image_id": "2112_N3YcO7YTOBE_00000483",
"label_matched": [
{
"mask_ids": [
6
],
"txt_desc": "man in dark shorts and a light gray full-sleeved pullover"
},
{
"mask_ids": [
2
],
"txt_desc": "upward-moving escalator"
},
{
"mask_ids": [
3
],
"txt_desc": "metal and glass railings"
},
{
"mask_ids": [
1
],
"txt_desc": "slatted, lighted ceiling"
},
{
"mask_ids": [
5
],
"txt_desc": "long fluorescent lamp and two identical round shaped lamps"
},
{
"mask_ids": [
4
],
"txt_desc": "directory sign reading \"2nd FLOOR\" and \"3rd FLOOR\""
},
{
"mask_ids": [
0
],
"txt_desc": "reddish-brown wall"
},
{
"mask_ids": [
7
],
"txt_desc": "partially visible person"
},
{
"mask_ids": [
6
],
"txt_desc": "man"
}
],
"labels": [
"wall",
"ceiling",
"escalator",
"handrail_or_fence",
"billboard_or_Bulletin_Board",
"lamp",
"person",
"person"
]
} | [
{
"area": 52566,
"bbox": [
0,
233,
1280,
362
],
"category_id": 0,
"id": 18440,
"image_id": "2112_N3YcO7YTOBE_00000483",
"iscrowd": 0,
"segmentation": {
"counts": "`7c2mc00000000000O10000O10000000000O10000O10000000000O10000O10000000000O10000O1000O100000O1... | [
{
"id": 1,
"name": "object"
}
] |
train | 2117_bOfZ7jc6hCI_00002418 | 2117_bOfZ7jc6hCI_00002418.jpg | {
"data_source": "VIPSeg",
"file_name": "2117_bOfZ7jc6hCI_00002418.jpg",
"height": 720,
"id": "2117_bOfZ7jc6hCI_00002418",
"width": 1280
} | {
"caption": "A man in a red long-sleeved shirt and grey sweatpants with a black jacket draped over his shoulder, and a brown bag across his body, is standing on an escalator to the left, looking towards a woman with long blonde hair who is wearing a red top and black shiny pants while descending the escalator on the right. A red bag is hanging from the man's side. A white wall is visible on the far left.",
"caption_ann": "A <3:man in a red long-sleeved shirt and grey sweatpants> with a <3:black jacket> draped over his shoulder, and a <4:brown bag> across his body, is standing on an <1:escalator> to the left, looking towards a <2:woman with long blonde hair> who is wearing a <2:red top and black shiny pants> while descending the <1:escalator> on the right. A <5:red bag> is hanging from the man's side. A <0:white wall> is visible on the far left.",
"id": 1702,
"image_id": "2117_bOfZ7jc6hCI_00002418",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "man in a red long-sleeved shirt and grey sweatpants"
},
{
"mask_ids": [
3
],
"txt_desc": "black jacket"
},
{
"mask_ids": [
4
],
"txt_desc": "brown bag"
},
{
"mask_ids": [
1
],
"txt_desc": "escalator"
},
{
"mask_ids": [
2
],
"txt_desc": "woman with long blonde hair"
},
{
"mask_ids": [
2
],
"txt_desc": "red top and black shiny pants"
},
{
"mask_ids": [
1
],
"txt_desc": "escalator"
},
{
"mask_ids": [
5
],
"txt_desc": "red bag"
},
{
"mask_ids": [
0
],
"txt_desc": "white wall"
}
],
"labels": [
"wall",
"escalator",
"person",
"person",
"bag_or_package",
"bag_or_package"
]
} | [
{
"area": 104632,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18448,
"image_id": "2117_bOfZ7jc6hCI_00002418",
"iscrowd": 0,
"segmentation": {
"counts": "0^gd12en[N;K5E;L4E;K5F:J6F:J6F:K5E;K5F:K5ZNgLj_Od3U`0]Lf_Oh3Y`0YL]_OQ4c`0nKX_OX4g`0i000O1K5E... | [
{
"id": 1,
"name": "object"
}
] |
train | 211_LqoV4ZW7xTA_00002211 | 211_LqoV4ZW7xTA_00002211.jpg | {
"data_source": "VIPSeg",
"file_name": "211_LqoV4ZW7xTA_00002211.jpg",
"height": 720,
"id": "211_LqoV4ZW7xTA_00002211",
"width": 1280
} | {
"caption": "A person in a dark suit sits on a black piano chair playing a black grand piano on a polished wooden floor.",
"caption_ann": "A <1:person in a dark suit> sits on a <2:black piano chair> playing a <3:black grand piano> on a <0:polished wooden floor>.",
"id": 1703,
"image_id": "211_LqoV4ZW7xTA_00002211",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "person in a dark suit"
},
{
"mask_ids": [
2
],
"txt_desc": "black piano chair"
},
{
"mask_ids": [
3
],
"txt_desc": "black grand piano"
},
{
"mask_ids": [
0
],
"txt_desc": "polished wooden floor"
}
],
"labels": [
"floor",
"person",
"chair_or_seat",
"instrument"
]
} | [
{
"area": 724271,
"bbox": [
0,
0,
1280,
720
],
"category_id": 13,
"id": 18454,
"image_id": "211_LqoV4ZW7xTA_00002211",
"iscrowd": 0,
"segmentation": {
"counts": "0_Xo6?\\]QI9G:J4M5K1O3M100O2N1O1O2N1N2O1N201L3O1K6K4jNUNZ]OX2cb0l0J6O1O1O1O101M20000O2N100O1... | [
{
"id": 1,
"name": "object"
}
] |
train | 2124_-YnYw0DjD1o_00000558 | 2124_-YnYw0DjD1o_00000558.jpg | {
"data_source": "VIPSeg",
"file_name": "2124_-YnYw0DjD1o_00000558.jpg",
"height": 720,
"id": "2124_-YnYw0DjD1o_00000558",
"width": 1280
} | {
"caption": "A person's hand holds a large green mango in the foreground, extended towards a young boy who is looking intently at the fruit. The boy is wearing a plaid shirt and is standing in a grassy area with dirt patches. To the left, a person wearing a green and white plaid shirt is partially visible. In the background, trees and wood can be seen under a blue sky.",
"caption_ann": "A <8:person's hand> holds a <5:large green mango> in the foreground, extended towards a <7:young boy> who is looking intently at the fruit. The boy is wearing a <7:plaid shirt> and is standing in a <1:grassy area> with <0:dirt patches>. To the left, a <6:person wearing a green and white plaid shirt> is partially visible. In the background, <4:trees> and <3:wood> can be seen under a <2:blue sky>.",
"id": 1704,
"image_id": "2124_-YnYw0DjD1o_00000558",
"label_matched": [
{
"mask_ids": [
8
],
"txt_desc": "person's hand"
},
{
"mask_ids": [
5
],
"txt_desc": "large green mango"
},
{
"mask_ids": [
7
],
"txt_desc": "young boy"
},
{
"mask_ids": [
7
],
"txt_desc": "plaid shirt"
},
{
"mask_ids": [
1
],
"txt_desc": "grassy area"
},
{
"mask_ids": [
0
],
"txt_desc": "dirt patches"
},
{
"mask_ids": [
6
],
"txt_desc": "person wearing a green and white plaid shirt"
},
{
"mask_ids": [
4
],
"txt_desc": "trees"
},
{
"mask_ids": [
3
],
"txt_desc": "wood"
},
{
"mask_ids": [
2
],
"txt_desc": "blue sky"
}
],
"labels": [
"ground",
"grass",
"sky",
"wood",
"tree",
"fruit",
"person",
"person",
"person"
]
} | [
{
"area": 43680,
"bbox": [
247,
0,
758,
213
],
"category_id": 14,
"id": 18458,
"image_id": "2124_-YnYw0DjD1o_00000558",
"iscrowd": 0,
"segmentation": {
"counts": "`i]53]f03L6K2N6J2N5K3L5L3M5J4M5K2N5K3L6K3M4K4M5K2N4L0O10000O10O10000000O01000O100000000O100... | [
{
"id": 1,
"name": "object"
}
] |
train | 2125_3lXlwRCjYLg_00001653 | 2125_3lXlwRCjYLg_00001653.jpg | {
"data_source": "VIPSeg",
"file_name": "2125_3lXlwRCjYLg_00001653.jpg",
"height": 720,
"id": "2125_3lXlwRCjYLg_00001653",
"width": 1280
} | {
"caption": "A person is slicing a dark green watermelon on an elongated wooden cutting board. Two slices of watermelon are already on the cutting board. The cutting board rests on a red table or desk. Towards the left side, part of a metal table is visible, with several other pieces of cut watermelon scattered on it. A yellow tool or sponge is also on the red table or desk to the right. A person in white pants stands behind on a dirt ground on the left and two other people are visible on the right. ",
"caption_ann": "A <3:person> is slicing a <2:dark green watermelon> on an <8:elongated wooden cutting board>. Two <2:slices of watermelon> are already on the cutting board. The cutting board rests on a <9:red table or desk>. Towards the left side, part of a <7:metal table> is visible, with <2:several other pieces of cut watermelon> scattered on it. A <1:yellow tool or sponge> is also on the <9:red table or desk> to the right. A <5:person in white pants> stands behind on a<0:dirt ground> on the left and <4,6:two other people> are visible on the right. ",
"id": 1705,
"image_id": "2125_3lXlwRCjYLg_00001653",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "person"
},
{
"mask_ids": [
2
],
"txt_desc": "dark green watermelon"
},
{
"mask_ids": [
8
],
"txt_desc": "elongated wooden cutting board"
},
{
"mask_ids": [
2
],
"txt_desc": "slices of watermelon"
},
{
"mask_ids": [
9
],
"txt_desc": "red table or desk"
},
{
"mask_ids": [
7
],
"txt_desc": "metal table"
},
{
"mask_ids": [
2
],
"txt_desc": "several other pieces of cut watermelon"
},
{
"mask_ids": [
1
],
"txt_desc": "yellow tool or sponge"
},
{
"mask_ids": [
9
],
"txt_desc": "red table or desk"
},
{
"mask_ids": [
5
],
"txt_desc": "person in white pants"
},
{
"mask_ids": [
0
],
"txt_desc": "dirt ground"
},
{
"mask_ids": [
4,
6
],
"txt_desc": "two other people"
}
],
"labels": [
"ground",
"tool",
"fruit",
"person",
"person",
"person",
"person",
"table_or_desk",
"table_or_desk",
"table_or_desk"
]
} | [
{
"area": 68314,
"bbox": [
910,
0,
370,
259
],
"category_id": 14,
"id": 18467,
"image_id": "2125_3lXlwRCjYLg_00001653",
"iscrowd": 0,
"segmentation": {
"counts": "WRPd04\\f06I7J6I5L0O100O1O100O100O1WOSOo[Om0Pd0\\Og[Oe0Xd0D`[O<_d0k0O100O1O1O1O1O1O1O1O1O10... | [
{
"id": 1,
"name": "object"
}
] |
train | 2126_6IjLT_eym0A_00002211 | 2126_6IjLT_eym0A_00002211.jpg | {
"data_source": "VIPSeg",
"file_name": "2126_6IjLT_eym0A_00002211.jpg",
"height": 720,
"id": "2126_6IjLT_eym0A_00002211",
"width": 1280
} | {
"caption": "A person's hand is reaching into a blue plastic basket filled with numerous round, purple, and green mangosteen fruits. The basket sits on the brown ground covered with green foliage and dried leaves.",
"caption_ann": "A <2:person's hand> is reaching into a <3:blue plastic basket> filled with <1:numerous round, purple, and green mangosteen fruits>. The basket sits on the <0:brown ground> covered with green foliage and dried leaves.",
"id": 1706,
"image_id": "2126_6IjLT_eym0A_00002211",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "person's hand"
},
{
"mask_ids": [
3
],
"txt_desc": "blue plastic basket"
},
{
"mask_ids": [
1
],
"txt_desc": "numerous round, purple, and green mangosteen fruits"
},
{
"mask_ids": [
0
],
"txt_desc": "brown ground"
}
],
"labels": [
"ground",
"fruit",
"person",
"basket"
]
} | [
{
"area": 169106,
"bbox": [
0,
0,
1280,
720
],
"category_id": 14,
"id": 18477,
"image_id": "2126_6IjLT_eym0A_00002211",
"iscrowd": 0,
"segmentation": {
"counts": "i>g7k>N002N002N001O1O001O1O1O1O1O2N1O2N1O1O2N1O1O1O1O1O1O1O1O1O1O3M100O1O3M1O2O0O2O0O2N1O1O... | [
{
"id": 1,
"name": "object"
}
] |
train | 2129_ZARE0K52e4w_00000288 | 2129_ZARE0K52e4w_00000288.jpg | {
"data_source": "VIPSeg",
"file_name": "2129_ZARE0K52e4w_00000288.jpg",
"height": 720,
"id": "2129_ZARE0K52e4w_00000288",
"width": 1280
} | {
"caption": "A man wearing a brown hat, a white shirt, and a dark vest is sitting on a wooden bench, holding a corn cob in his right hand. He is surrounded by tall green corn stalks. To the far left, a grey bench is partially visible on the gray cemented ground.",
"caption_ann": "A <3:man wearing a brown hat, a white shirt, and a dark vest> is sitting on a <5:wooden bench>, holding a <2:corn cob> in his right hand. He is surrounded by <1:tall green corn stalks>. To the far left, a <4:grey bench> is partially visible on the <0:gray cemented ground>.",
"id": 1707,
"image_id": "2129_ZARE0K52e4w_00000288",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "man wearing a brown hat, a white shirt, and a dark vest"
},
{
"mask_ids": [
5
],
"txt_desc": "wooden bench"
},
{
"mask_ids": [
2
],
"txt_desc": "corn cob"
},
{
"mask_ids": [
1
],
"txt_desc": "tall green corn stalks"
},
{
"mask_ids": [
4
],
"txt_desc": "grey bench"
},
{
"mask_ids": [
0
],
"txt_desc": "gray cemented ground"
}
],
"labels": [
"ground",
"other_plant",
"food",
"person",
"bench",
"bench"
]
} | [
{
"area": 10534,
"bbox": [
0,
519,
104,
140
],
"category_id": 14,
"id": 18481,
"image_id": "2129_ZARE0K52e4w_00000288",
"iscrowd": 0,
"segmentation": {
"counts": "W`0\\4Tb00000O1000000O1000000000000O10000O100000000000000O10000O9H8H000000001N1000001M2J6G9... | [
{
"id": 1,
"name": "object"
}
] |
train | 212_0vvyDjyJG18_00000772 | 212_0vvyDjyJG18_00000772.jpg | {
"data_source": "VIPSeg",
"file_name": "212_0vvyDjyJG18_00000772.jpg",
"height": 720,
"id": "212_0vvyDjyJG18_00000772",
"width": 1280
} | {
"caption": "A person in blue tshirt helps a person in wearing a green surgical gown, while another person in a green surgical gown stands nearby, all in a room with a light-colored wall adorned with a white posters with information written on them, a white board, a blue trash can, a green table, a surgical table, and a grey color printer.",
"caption_ann": "A <5:person in blue tshirt> helps a <3:person in wearing a green surgical gown>, while another <4:person in a green surgical gown> stands nearby, all in a room with a <0:light-colored wall> adorned with a <8,9,10:white posters with information written on them>, a <2:white board>, a <1:blue trash can>, a <6:green table>, a <7:surgical table>, and a <11:grey color printer>.",
"id": 1708,
"image_id": "212_0vvyDjyJG18_00000772",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "person in blue tshirt"
},
{
"mask_ids": [
3
],
"txt_desc": "person in wearing a green surgical gown"
},
{
"mask_ids": [
4
],
"txt_desc": "person in a green surgical gown"
},
{
"mask_ids": [
0
],
"txt_desc": "light-colored wall"
},
{
"mask_ids": [
8,
9,
10
],
"txt_desc": "white posters with information written on them"
},
{
"mask_ids": [
2
],
"txt_desc": "white board"
},
{
"mask_ids": [
1
],
"txt_desc": "blue trash can"
},
{
"mask_ids": [
6
],
"txt_desc": "green table"
},
{
"mask_ids": [
7
],
"txt_desc": "surgical table"
},
{
"mask_ids": [
11
],
"txt_desc": "grey color printer"
}
],
"labels": [
"wall",
"trash_can",
"blackboard",
"person",
"person",
"person",
"table_or_desk",
"table_or_desk",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"printer"
]
} | [
{
"area": 259575,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18487,
"image_id": "212_0vvyDjyJG18_00000772",
"iscrowd": 0,
"segmentation": {
"counts": "0l4da01O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 212_OroNM1E-A9c_00000486 | 212_OroNM1E-A9c_00000486.jpg | {
"data_source": "VIPSeg",
"file_name": "212_OroNM1E-A9c_00000486.jpg",
"height": 720,
"id": "212_OroNM1E-A9c_00000486",
"width": 1280
} | {
"caption": "A man in a tan suit jacket sits at a black upright piano with its top open, in a room with a white wall that has a white shelf above the piano and a small framed picture hanging on the wall.",
"caption_ann": "A <2:man in a tan suit jacket> sits at a <4:black upright piano> with its top open, in a room with a <0:white wall> that has a <1:white shelf> above the <4:piano> and a <3:small framed picture> hanging on the <0:wall>.",
"id": 1709,
"image_id": "212_OroNM1E-A9c_00000486",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "man in a tan suit jacket"
},
{
"mask_ids": [
4
],
"txt_desc": "black upright piano"
},
{
"mask_ids": [
0
],
"txt_desc": "white wall"
},
{
"mask_ids": [
1
],
"txt_desc": "white shelf"
},
{
"mask_ids": [
4
],
"txt_desc": "piano"
},
{
"mask_ids": [
3
],
"txt_desc": "small framed picture"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
}
],
"labels": [
"wall",
"shelf",
"person",
"painting_or_poster",
"instrument"
]
} | [
{
"area": 276436,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18499,
"image_id": "212_OroNM1E-A9c_00000486",
"iscrowd": 0,
"segmentation": {
"counts": "0`S21obN00000000000000000000000000000000000000001O000000000000000000000\\A1g6OPI<n6DoH`0P7@oH... | [
{
"id": 1,
"name": "object"
}
] |
train | 2130_k4ZP8Kxp7Qc_00000393 | 2130_k4ZP8Kxp7Qc_00000393.jpg | {
"data_source": "VIPSeg",
"file_name": "2130_k4ZP8Kxp7Qc_00000393.jpg",
"height": 720,
"id": "2130_k4ZP8Kxp7Qc_00000393",
"width": 1280
} | {
"caption": "A woman with dark hair tied in a ponytail is wearing a red shirt and reaching up to touch a pink dragon fruit growing amidst lush green cactus-like plants. The blue sky is visible in the upper portion of the image.",
"caption_ann": "A <3:woman with dark hair tied in a ponytail> is wearing a <3:red shirt> and reaching up to touch a <2:pink dragon fruit> growing amidst <1:lush green cactus-like plants>. The <0:blue sky> is visible in the upper portion of the image.",
"id": 1710,
"image_id": "2130_k4ZP8Kxp7Qc_00000393",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "woman with dark hair tied in a ponytail"
},
{
"mask_ids": [
3
],
"txt_desc": "red shirt"
},
{
"mask_ids": [
2
],
"txt_desc": "pink dragon fruit"
},
{
"mask_ids": [
1
],
"txt_desc": "lush green cactus-like plants"
},
{
"mask_ids": [
0
],
"txt_desc": "blue sky"
}
],
"labels": [
"sky",
"other_plant",
"fruit",
"person"
]
} | [
{
"area": 102479,
"bbox": [
0,
0,
1280,
191
],
"category_id": 28,
"id": 18504,
"image_id": "2130_k4ZP8Kxp7Qc_00000393",
"iscrowd": 0,
"segmentation": {
"counts": "0^1?_O4Gla0i0Q^O_O4Hka0i0Q^O_O4Hka0i0Q^O_O4Hka0i0Q^O_O4Hka0i0Q^O_O4Hka0i0Q^O_O4Ija0h0R^O^O4... | [
{
"id": 1,
"name": "object"
}
] |
train | 2131_l5m4Y9FCKwo_00002178 | 2131_l5m4Y9FCKwo_00002178.jpg | {
"data_source": "VIPSeg",
"file_name": "2131_l5m4Y9FCKwo_00002178.jpg",
"height": 720,
"id": "2131_l5m4Y9FCKwo_00002178",
"width": 1280
} | {
"caption": "A topless man wearing olive green shorts is sitting on grey rocks in a lush, green environment with trees and foliage. He is holding a red apple in each hand, and three more red apples are on the rocks in front of him. A large green and yellow leaf is prominent in the foreground on the left.",
"caption_ann": "A <3:topless man wearing olive green shorts> is sitting on <0:grey rocks> in a lush, green environment with <1:trees and foliage>. He is holding a <2:red apple> in each hand, and <2:three more red apples> are on the rocks in front of him. A <1:large green and yellow leaf> is prominent in the foreground on the left.",
"id": 1711,
"image_id": "2131_l5m4Y9FCKwo_00002178",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "topless man wearing olive green shorts"
},
{
"mask_ids": [
0
],
"txt_desc": "grey rocks"
},
{
"mask_ids": [
1
],
"txt_desc": "trees and foliage"
},
{
"mask_ids": [
2
],
"txt_desc": "red apple"
},
{
"mask_ids": [
2
],
"txt_desc": "three more red apples"
},
{
"mask_ids": [
1
],
"txt_desc": "large green and yellow leaf"
}
],
"labels": [
"stone",
"tree",
"fruit",
"person"
]
} | [
{
"area": 141770,
"bbox": [
277,
181,
958,
539
],
"category_id": 30,
"id": 18508,
"image_id": "2131_l5m4Y9FCKwo_00002178",
"iscrowd": 0,
"segmentation": {
"counts": "o^S61]f02M3N2M3N2L4N2L4N2M3N2L4O1L4N2L4O1O1N2O1O1O1O1N200N200N2O1O1O1N2O1O1O1N200N200N2O... | [
{
"id": 1,
"name": "object"
}
] |
train | 2143_6OMR3X7IcZ0_00002433 | 2143_6OMR3X7IcZ0_00002433.jpg | {
"data_source": "VIPSeg",
"file_name": "2143_6OMR3X7IcZ0_00002433.jpg",
"height": 720,
"id": "2143_6OMR3X7IcZ0_00002433",
"width": 1280
} | {
"caption": "A person's hand with a red bracelet on the wrist is placing a flat, round piece of food into a stainless steel microwave oven. The microwave has a digital display showing \"1:33\" and is sitting on a dark granite countertop. A white wall is visible to the right.",
"caption_ann": "A <2:person's hand> with a <2:red bracelet> on the wrist is placing a <1:flat, round piece of food> into a <4:stainless steel microwave oven>. The microwave has a <4:digital display showing \"1:33\"> and is sitting on a <3:dark granite countertop>. A <0:white wall> is visible to the right.",
"id": 1712,
"image_id": "2143_6OMR3X7IcZ0_00002433",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "person's hand"
},
{
"mask_ids": [
2
],
"txt_desc": "red bracelet"
},
{
"mask_ids": [
1
],
"txt_desc": "flat, round piece of food"
},
{
"mask_ids": [
4
],
"txt_desc": "stainless steel microwave oven"
},
{
"mask_ids": [
4
],
"txt_desc": "digital display showing \"1:33\""
},
{
"mask_ids": [
3
],
"txt_desc": "dark granite countertop"
},
{
"mask_ids": [
0
],
"txt_desc": "white wall"
}
],
"labels": [
"wall",
"food",
"person",
"table_or_desk",
"Microwave_oven"
]
} | [
{
"area": 31438,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18512,
"image_id": "2143_6OMR3X7IcZ0_00002433",
"iscrowd": 0,
"segmentation": {
"counts": "0]1Se00000000000000000000000000000000000000000000000000000000000000000000000O10000O1000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 2149_lInFVrlnzvc_00002448 | 2149_lInFVrlnzvc_00002448.jpg | {
"data_source": "VIPSeg",
"file_name": "2149_lInFVrlnzvc_00002448.jpg",
"height": 720,
"id": "2149_lInFVrlnzvc_00002448",
"width": 1280
} | {
"caption": "A person is seen using a white cloth to remove a small clear glass bowl from inside a stainless steel microwave oven. The microwave is situated on a dark countertop, and a tiled wall is visible in the background on the left.",
"caption_ann": "A <2:person> is seen using a <1:white cloth> to remove a <3:small clear glass bowl> from inside a <5:stainless steel microwave oven>. The microwave is situated on a <4:dark countertop>, and a <0:tiled wall> is visible in the background on the left.",
"id": 1713,
"image_id": "2149_lInFVrlnzvc_00002448",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "person"
},
{
"mask_ids": [
1
],
"txt_desc": "white cloth"
},
{
"mask_ids": [
3
],
"txt_desc": "small clear glass bowl"
},
{
"mask_ids": [
5
],
"txt_desc": "stainless steel microwave oven"
},
{
"mask_ids": [
4
],
"txt_desc": "dark countertop"
},
{
"mask_ids": [
0
],
"txt_desc": "tiled wall"
}
],
"labels": [
"wall",
"textiles",
"person",
"tub_or_bowl_or_pot",
"table_or_desk",
"Microwave_oven"
]
} | [
{
"area": 168530,
"bbox": [
0,
0,
1280,
392
],
"category_id": 0,
"id": 18517,
"image_id": "2149_lInFVrlnzvc_00002448",
"iscrowd": 0,
"segmentation": {
"counts": "0U<[:1O0000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 2151_0RZ_en3H1gQ_00003306 | 2151_0RZ_en3H1gQ_00003306.jpg | {
"data_source": "VIPSeg",
"file_name": "2151_0RZ_en3H1gQ_00003306.jpg",
"height": 720,
"id": "2151_0RZ_en3H1gQ_00003306",
"width": 1280
} | {
"caption": "A young girl with dark hair in a ponytail is sitting on a multi-coloured inflatable cushion with her legs extended, wearing pink socks. The background has a dark brown fence adjacent to the grass. A small tree is also visible to the right side of the cushion.",
"caption_ann": "A <4:young girl with dark hair in a ponytail> is sitting on a <2:multi-coloured inflatable cushion> with her legs extended, wearing pink socks. The background has a <0:dark brown fence> adjacent to the <1:grass>. A <3:small tree> is also visible to the right side of the cushion.",
"id": 1714,
"image_id": "2151_0RZ_en3H1gQ_00003306",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "young girl with dark hair in a ponytail"
},
{
"mask_ids": [
2
],
"txt_desc": "multi-coloured inflatable cushion"
},
{
"mask_ids": [
0
],
"txt_desc": "dark brown fence"
},
{
"mask_ids": [
1
],
"txt_desc": "grass"
},
{
"mask_ids": [
3
],
"txt_desc": "small tree"
}
],
"labels": [
"handrail_or_fence",
"grass",
"cushion_or_carpet",
"tree",
"person"
]
} | [
{
"area": 47915,
"bbox": [
0,
0,
944,
84
],
"category_id": 7,
"id": 18523,
"image_id": "2151_0RZ_en3H1gQ_00003306",
"iscrowd": 0,
"segmentation": {
"counts": "0c2mc000001O00000000000000000000O10000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 2152_5O4h5zzraN8_00003213 | 2152_5O4h5zzraN8_00003213.jpg | {
"data_source": "VIPSeg",
"file_name": "2152_5O4h5zzraN8_00003213.jpg",
"height": 720,
"id": "2152_5O4h5zzraN8_00003213",
"width": 1280
} | {
"caption": "A young boy in a dark t-shirt stands with his back to the camera on a sandy playground, with a yellow and gray climbing structure partially visible in the foreground. Behind the boy is a low stone wall that separates the sandy area from a grassy patch where several trees and thin poles are planted. In the background, beyond the grass, a light-colored building is visible, along with multiple parked cars on what appears to be a street under a bright sky.",
"caption_ann": "A <15:young boy in a dark t-shirt> stands with his back to the camera on a <2:sandy playground>, with a <4:yellow and gray climbing structure> partially visible in the foreground. Behind the boy is a <6:low stone wall> that separates the sandy area from a <1:grassy patch> where several <7:trees> and <0:thin poles> are planted. In the background, beyond the <1:grass>, a <3:light-colored building> is visible, along with multiple <8,9,10,11,12,13,14:parked cars> on what appears to be a <5:street> under a bright <5:sky>.",
"id": 1715,
"image_id": "2152_5O4h5zzraN8_00003213",
"label_matched": [
{
"mask_ids": [
15
],
"txt_desc": "young boy in a dark t-shirt"
},
{
"mask_ids": [
2
],
"txt_desc": "sandy playground"
},
{
"mask_ids": [
4
],
"txt_desc": "yellow and gray climbing structure"
},
{
"mask_ids": [
6
],
"txt_desc": "low stone wall"
},
{
"mask_ids": [
1
],
"txt_desc": "grassy patch"
},
{
"mask_ids": [
7
],
"txt_desc": "trees"
},
{
"mask_ids": [
0
],
"txt_desc": "thin poles"
},
{
"mask_ids": [
1
],
"txt_desc": "grass"
},
{
"mask_ids": [
3
],
"txt_desc": "light-colored building"
},
{
"mask_ids": [
8,
9,
10,
11,
12,
13,
14
],
"txt_desc": "parked cars"
},
{
"mask_ids": [
5
],
"txt_desc": "street"
},
{
"mask_ids": [
5
],
"txt_desc": "sky"
}
],
"labels": [
"pole",
"grass",
"sand",
"house",
"other_construction",
"sky",
"stone",
"tree",
"car",
"car",
"car",
"car",
"car",
"car",
"car",
"person"
]
} | [
{
"area": 4667,
"bbox": [
67,
0,
893,
210
],
"category_id": 12,
"id": 18528,
"image_id": "2152_5O4h5zzraN8_00003213",
"iscrowd": 0,
"segmentation": {
"counts": "]W_1f1fc0T1iMW2^Ob000000000000000000lMT2dNeRcb0VN[P]]O6C<J7D<I6D=J5E<I7C<K6C<K6C34NKO8OJO8OJO... | [
{
"id": 1,
"name": "object"
}
] |
train | 2154_8XB_0x_erho_00000561 | 2154_8XB_0x_erho_00000561.jpg | {
"data_source": "VIPSeg",
"file_name": "2154_8XB_0x_erho_00000561.jpg",
"height": 720,
"id": "2154_8XB_0x_erho_00000561",
"width": 1280
} | {
"caption": "A small child dressed in a green jacket and brown pants is seen at the bottom of a bright yellow spiral slide on a playground with dark gray rubberized ground tiles. In the background, there are various trees and other plants, along with a large rock formation and a building. Other playground equipment, including a red teeter-totter and a green bench, are also visible. On the left side of the image, two parked cars and a person are partially obscured by foliage.",
"caption_ann": "A <9:small child dressed in a green jacket and brown pants> is seen at the bottom of a <0:bright yellow spiral slide> on a <1:playground with dark gray rubberized ground tiles>. In the background, there are various <4:trees> and <5:other plants>, along with a <3:large rock formation> and a <2:building>. Other playground equipment, including a <6:red teeter-totter> and a <11:green bench>, are also visible. On the left side of the image, two <7,8:parked cars> and a <10:person> are partially obscured by foliage.",
"id": 1716,
"image_id": "2154_8XB_0x_erho_00000561",
"label_matched": [
{
"mask_ids": [
9
],
"txt_desc": "small child dressed in a green jacket and brown pants"
},
{
"mask_ids": [
0
],
"txt_desc": "bright yellow spiral slide"
},
{
"mask_ids": [
1
],
"txt_desc": "playground with dark gray rubberized ground tiles"
},
{
"mask_ids": [
4
],
"txt_desc": "trees"
},
{
"mask_ids": [
5
],
"txt_desc": "other plants"
},
{
"mask_ids": [
3
],
"txt_desc": "large rock formation"
},
{
"mask_ids": [
2
],
"txt_desc": "building"
},
{
"mask_ids": [
6
],
"txt_desc": "red teeter-totter"
},
{
"mask_ids": [
11
],
"txt_desc": "green bench"
},
{
"mask_ids": [
7,
8
],
"txt_desc": "parked cars"
},
{
"mask_ids": [
10
],
"txt_desc": "person"
}
],
"labels": [
"Playground_slide",
"ground",
"building",
"stone",
"tree",
"other_plant",
"toy",
"car",
"car",
"person",
"person",
"bench"
]
} | [
{
"area": 220942,
"bbox": [
354,
0,
553,
698
],
"category_id": 6,
"id": 18544,
"image_id": "2154_8XB_0x_erho_00000561",
"iscrowd": 0,
"segmentation": {
"counts": "eUi78je0c0H>A8J6I:G6K7I2O5J4L5K4L5K3M4L3K5M3M3M3M3N1M4N1N3M2N3M2O1O1M4N2M2N2O2M2N2O1M3O1N2N... | [
{
"id": 1,
"name": "object"
}
] |
train | 2156_gikFqhScCvU_00001698 | 2156_gikFqhScCvU_00001698.jpg | {
"data_source": "VIPSeg",
"file_name": "2156_gikFqhScCvU_00001698.jpg",
"height": 720,
"id": "2156_gikFqhScCvU_00001698",
"width": 1280
} | {
"caption": "A young, shirtless boy is sliding down a yellow water slide adorned with colorful footprints, heading towards a splash into the blue water below. The slide is part of a larger water park structure that features a blue wall with yellow accents and a red section on the left, from which water is actively cascading.",
"caption_ann": "A <3:young, shirtless boy> is sliding down a <0:yellow water slide adorned with colorful footprints>, heading towards a splash into the <2:blue water> below. The slide is part of a larger <1:water park structure> that features a blue wall with yellow accents and a red section on the left, from which water is actively cascading.",
"id": 1717,
"image_id": "2156_gikFqhScCvU_00001698",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "young, shirtless boy"
},
{
"mask_ids": [
0
],
"txt_desc": "yellow water slide adorned with colorful footprints"
},
{
"mask_ids": [
2
],
"txt_desc": "blue water"
},
{
"mask_ids": [
1
],
"txt_desc": "water park structure"
}
],
"labels": [
"Playground_slide",
"other_construction",
"water",
"person"
]
} | [
{
"area": 234553,
"bbox": [
467,
0,
429,
720
],
"category_id": 6,
"id": 18556,
"image_id": "2156_gikFqhScCvU_00001698",
"iscrowd": 0,
"segmentation": {
"counts": "iQY:7Uf04G9K5H8K5G9K5G9L4F:L4G9K5H8K5G9L4F:L4G9K5G9K5H8K5G9L4F:L4G9K5G9K5H8K5G9L4F:L4G9K5G9... | [
{
"id": 1,
"name": "object"
}
] |
train | 2157_jS6kQVDQmgw_00000828 | 2157_jS6kQVDQmgw_00000828.jpg | {
"data_source": "VIPSeg",
"file_name": "2157_jS6kQVDQmgw_00000828.jpg",
"height": 720,
"id": "2157_jS6kQVDQmgw_00000828",
"width": 1280
} | {
"caption": "A woman wearing glasses and a dark hoodie with white striped leggings walks a small, fluffy dog in a pink sweater on a leash across a grassy area. In the background, there is a white railing separating the grass from a paved area where numerous cars are parked. A tree trunk is visible to the right of the woman, and parts of some buildings and the sky can be seen in the upper background.",
"caption_ann": "A <16:woman wearing glasses and a dark hoodie with white striped leggings> walks a <17:small, fluffy dog in a pink sweater> on a leash across a <2:grassy area>. In the background, there is a <0:white railing> separating the <2:grass> from a <1:paved area> where numerous <6,7,8,9,10,11,12,13,14,15:cars> are parked. A <5:tree trunk> is visible to the right of the <16:woman>, and parts of some <3:buildings> and the <4:sky> can be seen in the upper background.",
"id": 1718,
"image_id": "2157_jS6kQVDQmgw_00000828",
"label_matched": [
{
"mask_ids": [
16
],
"txt_desc": "woman wearing glasses and a dark hoodie with white striped leggings"
},
{
"mask_ids": [
17
],
"txt_desc": "small, fluffy dog in a pink sweater"
},
{
"mask_ids": [
2
],
"txt_desc": "grassy area"
},
{
"mask_ids": [
0
],
"txt_desc": "white railing"
},
{
"mask_ids": [
2
],
"txt_desc": "grass"
},
{
"mask_ids": [
1
],
"txt_desc": "paved area"
},
{
"mask_ids": [
6,
7,
8,
9,
10,
11,
12,
13,
14,
15
],
"txt_desc": "cars"
},
{
"mask_ids": [
5
],
"txt_desc": "tree trunk"
},
{
"mask_ids": [
16
],
"txt_desc": "woman"
},
{
"mask_ids": [
3
],
"txt_desc": "buildings"
},
{
"mask_ids": [
4
],
"txt_desc": "sky"
}
],
"labels": [
"handrail_or_fence",
"ground",
"grass",
"house",
"sky",
"tree",
"car",
"car",
"car",
"car",
"car",
"car",
"car",
"car",
"car",
"car",
"person",
"dog"
]
} | [
{
"area": 18657,
"bbox": [
0,
162,
822,
70
],
"category_id": 7,
"id": 18560,
"image_id": "2157_jS6kQVDQmgw_00000828",
"iscrowd": 0,
"segmentation": {
"counts": "U5`0Pf0000000000000000000001O0000000000000000001N100000000000000000000O100000000000000000001O... | [
{
"id": 1,
"name": "object"
}
] |
train | 2159_I3Vf_ylAeOs_00006096 | 2159_I3Vf_ylAeOs_00006096.jpg | {
"data_source": "VIPSeg",
"file_name": "2159_I3Vf_ylAeOs_00006096.jpg",
"height": 720,
"id": "2159_I3Vf_ylAeOs_00006096",
"width": 1280
} | {
"caption": "On a perfectly manicured, vibrant green soccer pitch, a player in a dark uniform has just unleashed a powerful shot, sending the soccer ball flying towards the goal with a white net. Reacting with incredible agility, the goalkeeper, clad in a bright yellow kit, executes a full-stretch dive in a desperate attempt to block the shot. The scene is filled with anticipation as other players, including a defender in a white uniform, another player in a dark kit, and two opposing players in red jerseys numbered 10 and 32 respectively, watch the critical play unfold before the digital advertising billboards lining the side of the field.",
"caption_ann": "On a <0:perfectly manicured, vibrant green soccer pitch>, a <5:player in a dark uniform> has just unleashed a powerful shot, sending the <9,10:soccer ball> flying towards the <2:goal with a white net>. Reacting with incredible agility, the <3:goalkeeper, clad in a bright yellow kit,> executes a full-stretch dive in a desperate attempt to block the shot. The scene is filled with anticipation as other players, including a <4:defender in a white uniform>, another <6:player in a dark kit>, and <7,8:two opposing players in red jerseys numbered 10 and 32 respectively>, watch the critical play unfold before the <1:digital advertising billboards> lining the side of the field.",
"id": 1719,
"image_id": "2159_I3Vf_ylAeOs_00006096",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "perfectly manicured, vibrant green soccer pitch"
},
{
"mask_ids": [
5
],
"txt_desc": "player in a dark uniform"
},
{
"mask_ids": [
9,
10
],
"txt_desc": "soccer ball"
},
{
"mask_ids": [
2
],
"txt_desc": "goal with a white net"
},
{
"mask_ids": [
3
],
"txt_desc": "goalkeeper, clad in a bright yellow kit,"
},
{
"mask_ids": [
4
],
"txt_desc": "defender in a white uniform"
},
{
"mask_ids": [
6
],
"txt_desc": "player in a dark kit"
},
{
"mask_ids": [
7,
8
],
"txt_desc": "two opposing players in red jerseys numbered 10 and 32 respectively"
},
{
"mask_ids": [
1
],
"txt_desc": "digital advertising billboards"
}
],
"labels": [
"grass",
"billboard_or_Bulletin_Board",
"goal",
"person",
"person",
"person",
"person",
"person",
"person",
"ball",
"ball"
]
} | [
{
"area": 801117,
"bbox": [
0,
0,
1280,
720
],
"category_id": 15,
"id": 18578,
"image_id": "2159_I3Vf_ylAeOs_00006096",
"iscrowd": 0,
"segmentation": {
"counts": "0Ymb01Pi]O;I2N4L101O00000100O1O010N20O01O10000O1000O1000OQNLd]O4[b0Nc]O3Zb03c]OM\\b06b]OJ\\... | [
{
"id": 1,
"name": "object"
}
] |
train | 2168_ADEh9F0Lidc_00001518 | 2168_ADEh9F0Lidc_00001518.jpg | {
"data_source": "VIPSeg",
"file_name": "2168_ADEh9F0Lidc_00001518.jpg",
"height": 720,
"id": "2168_ADEh9F0Lidc_00001518",
"width": 1280
} | {
"caption": "A man in a blue t-shirt with a white watch sits in a blue chair at a dark desk, gesturing with his hands in front of an illuminated mechanical keyboard. Behind him, against a light gray wall, is a large flat-screen television displaying a mountain landscape with a sunset, resting on a dark console table. To the right of the man, a modern light fixture with two glowing vertical bars stands beside a dark computer tower.",
"caption_ann": "A <2:man in a blue t-shirt with a white watch> sits in a <5:blue chair> at a <3:dark desk>, gesturing with his hands in front of an <8:illuminated mechanical keyboard>. Behind him, against a <0:light gray wall>, is a <6:large flat-screen television> displaying a mountain landscape with a sunset, resting on a <4:dark console table>. To the right of the <2:man>, a <1:modern light fixture> with two glowing vertical bars stands beside a <7:dark computer tower>.",
"id": 1720,
"image_id": "2168_ADEh9F0Lidc_00001518",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "man in a blue t-shirt with a white watch"
},
{
"mask_ids": [
5
],
"txt_desc": "blue chair"
},
{
"mask_ids": [
3
],
"txt_desc": "dark desk"
},
{
"mask_ids": [
8
],
"txt_desc": "illuminated mechanical keyboard"
},
{
"mask_ids": [
0
],
"txt_desc": "light gray wall"
},
{
"mask_ids": [
6
],
"txt_desc": "large flat-screen television"
},
{
"mask_ids": [
4
],
"txt_desc": "dark console table"
},
{
"mask_ids": [
2
],
"txt_desc": "man"
},
{
"mask_ids": [
1
],
"txt_desc": "modern light fixture"
},
{
"mask_ids": [
7
],
"txt_desc": "dark computer tower"
}
],
"labels": [
"wall",
"lamp",
"person",
"table_or_desk",
"table_or_desk",
"chair_or_seat",
"screen_or_television",
"computer",
"keyboard"
]
} | [
{
"area": 412844,
"bbox": [
0,
0,
1280,
571
],
"category_id": 0,
"id": 18589,
"image_id": "2168_ADEh9F0Lidc_00001518",
"iscrowd": 0,
"segmentation": {
"counts": "0aa0o400000000000000000000000000000000000000XMh200O10000N20000O10000O10000O100O100O10000O100... | [
{
"id": 1,
"name": "object"
}
] |
train | 2169_bPWl9Apq9hk_00002334 | 2169_bPWl9Apq9hk_00002334.jpg | {
"data_source": "VIPSeg",
"file_name": "2169_bPWl9Apq9hk_00002334.jpg",
"height": 720,
"id": "2169_bPWl9Apq9hk_00002334",
"width": 1280
} | {
"caption": "A person is actively engaged at a dark glossy desk, with their left hand on a black computer mouse and their right hand typing on a black keyboard with red backlighting presumibly looking at a screen with only the support visible. The desk features a large mouse pad with red accents and a white logo. The person is seated on a red and black gaming chair. In the background, another matching gaming chair is visible and an additional keyboard, along with parts of a wall and wooden floor.",
"caption_ann": "A <4:person> is actively engaged at a <5:dark glossy desk>, with their left hand on a <3:black computer mouse> and their right hand typing on a <9:black keyboard with red backlighting> presumibly looking at a <8:screen with only the support visible>. The <5:desk> features a <2:large mouse pad with red accents and a white logo>. The person is seated on a <6:red and black gaming chair>. In the background, <7:another matching gaming chair> is visible and an <10:additional keyboard>, along with parts of a <0:wall> and <1:wooden floor>.",
"id": 1721,
"image_id": "2169_bPWl9Apq9hk_00002334",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "person"
},
{
"mask_ids": [
5
],
"txt_desc": "dark glossy desk"
},
{
"mask_ids": [
3
],
"txt_desc": "black computer mouse"
},
{
"mask_ids": [
9
],
"txt_desc": "black keyboard with red backlighting"
},
{
"mask_ids": [
8
],
"txt_desc": "screen with only the support visible"
},
{
"mask_ids": [
5
],
"txt_desc": "desk"
},
{
"mask_ids": [
2
],
"txt_desc": "large mouse pad with red accents and a white logo"
},
{
"mask_ids": [
6
],
"txt_desc": "red and black gaming chair"
},
{
"mask_ids": [
7
],
"txt_desc": "another matching gaming chair"
},
{
"mask_ids": [
10
],
"txt_desc": "additional keyboard"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
},
{
"mask_ids": [
1
],
"txt_desc": "wooden floor"
}
],
"labels": [
"wall",
"floor",
"cushion_or_carpet",
"other_electronic_product",
"person",
"table_or_desk",
"chair_or_seat",
"chair_or_seat",
"screen_or_television",
"keyboard",
"keyboard"
]
} | [
{
"area": 10538,
"bbox": [
0,
12,
632,
261
],
"category_id": 0,
"id": 18598,
"image_id": "2169_bPWl9Apq9hk_00002334",
"iscrowd": 0,
"segmentation": {
"counts": "m0d7n>N1O2N1O2M102M2O3M004L1N4L3N2N1O3L2O3L2O3M2N4K2O3L2O3M2N4K3M6K2N3L3M5L2N4K3M4M2N3L3N4K2O... | [
{
"id": 1,
"name": "object"
}
] |
train | 216_hSIYGZhRGd4_00000961 | 216_hSIYGZhRGd4_00000961.jpg | {
"data_source": "VIPSeg",
"file_name": "216_hSIYGZhRGd4_00000961.jpg",
"height": 720,
"id": "216_hSIYGZhRGd4_00000961",
"width": 1280
} | {
"caption": "A bustling scene unfolds on a paved path in a lush green park. In the middle ground, a street artist has set up a stall under a large blue parasol. A woman and a artist sit on a folding chair, attending to a display of numerous black and white portrait eleven drawings. Nearby, a man relaxes in a bench with his hands behind his head.five pedestrians are walking along the path. On the left, a person in a light-colored shirt and khaki shorts carries a dark bag, and on the right, a man wearing a green shirt walks alongside a man in a white top. In the background, a long row of dark park four benches sits behind a fence, all under the canopy of leafy trees and a patch of sky.",
"caption_ann": "A bustling scene unfolds on a <1:paved path> in a lush green park. In the middle ground, a street <6:artist> has set up a stall under a large blue <4:parasol>. A <5:woman> and a <6:artist> sit on a <14:folding chair>, attending to a display of numerous black and white portrait <19,20,21,22,23,24,25,26,27,28,29:eleven drawings>. Nearby, a <7:man> relaxes in a <16:bench> with his hands behind his head.<8,9,10,11,12:five pedestrians> are walking along the path. On the left, a <8:person> in a light-colored shirt and khaki shorts carries a <13:dark bag>, and on the right, a <11:man wearing a green shirt> walks alongside a <10,12:man in a white top>. In the background, a long row of dark park <15,16,17,18:four benches> sits behind a <0:fence>, all under the canopy of <3:leafy trees> and a patch of <2:sky>.",
"id": 1722,
"image_id": "216_hSIYGZhRGd4_00000961",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "paved path"
},
{
"mask_ids": [
6
],
"txt_desc": "artist"
},
{
"mask_ids": [
4
],
"txt_desc": "parasol"
},
{
"mask_ids": [
5
],
"txt_desc": "woman"
},
{
"mask_ids": [
6
],
"txt_desc": "artist"
},
{
"mask_ids": [
14
],
"txt_desc": "folding chair"
},
{
"mask_ids": [
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29
],
"txt_desc": "eleven drawings"
},
{
"mask_ids": [
7
],
"txt_desc": "man"
},
{
"mask_ids": [
16
],
"txt_desc": "bench"
},
{
"mask_ids": [
8,
9,
10,
11,
12
],
"txt_desc": "five pedestrians"
},
{
"mask_ids": [
8
],
"txt_desc": "person"
},
{
"mask_ids": [
13
],
"txt_desc": "dark bag"
},
{
"mask_ids": [
11
],
"txt_desc": "man wearing a green shirt"
},
{
"mask_ids": [
10,
12
],
"txt_desc": "man in a white top"
},
{
"mask_ids": [
15,
16,
17,
18
],
"txt_desc": "four benches"
},
{
"mask_ids": [
0
],
"txt_desc": "fence"
},
{
"mask_ids": [
3
],
"txt_desc": "leafy trees"
},
{
"mask_ids": [
2
],
"txt_desc": "sky"
}
],
"labels": [
"handrail_or_fence",
"ground",
"sky",
"tree",
"parasol_or_umbrella",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"bag_or_package",
"chair_or_seat",
"bench",
"bench",
"bench",
"bench",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster"
]
} | [
{
"area": 56304,
"bbox": [
0,
384,
1188,
152
],
"category_id": 7,
"id": 18609,
"image_id": "216_hSIYGZhRGd4_00000961",
"iscrowd": 0,
"segmentation": {
"counts": "e<h0Q3B\\?>c@B^?>b@B^?>b@B^?>a@C_?=a@C_?=`@D`?<`@D`?<`@D`?<m^OZOoNa1Tb0UOj^O^OQO]1Ub0UOf^OBT... | [
{
"id": 1,
"name": "object"
}
] |
train | 216_ln63SOGqztc_00001968 | 216_ln63SOGqztc_00001968.jpg | {
"data_source": "VIPSeg",
"file_name": "216_ln63SOGqztc_00001968.jpg",
"height": 720,
"id": "216_ln63SOGqztc_00001968",
"width": 1280
} | {
"caption": "On a stage with a wooden floor, a person wearing a dark vest over a white shirt performs a piano recital. The performer is sits on a dark bench to play a black upright piano in front of a deep blue, pleated curtain. To the right of the stage, a Texas state flag is displayed against a light brown-colored brick wall. A person's head visible in the bottom right corner.",
"caption_ann": "On a stage with a <1:wooden floor>, a <4:person wearing a dark vest over a white shirt> performs a piano recital. The <4:performer> is sits on a <6:dark bench> to play a <7:black upright piano> in front of a <2:deep blue, pleated curtain>. To the right of the stage, a <3:Texas state flag> is displayed against a <0:light brown-colored brick wall>. A <5:person's head> visible in the bottom right corner.",
"id": 1723,
"image_id": "216_ln63SOGqztc_00001968",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "wooden floor"
},
{
"mask_ids": [
4
],
"txt_desc": "person wearing a dark vest over a white shirt"
},
{
"mask_ids": [
4
],
"txt_desc": "performer"
},
{
"mask_ids": [
6
],
"txt_desc": "dark bench"
},
{
"mask_ids": [
7
],
"txt_desc": "black upright piano"
},
{
"mask_ids": [
2
],
"txt_desc": "deep blue, pleated curtain"
},
{
"mask_ids": [
3
],
"txt_desc": "Texas state flag"
},
{
"mask_ids": [
0
],
"txt_desc": "light brown-colored brick wall"
},
{
"mask_ids": [
5
],
"txt_desc": "person's head"
}
],
"labels": [
"wall",
"floor",
"textiles",
"flag",
"person",
"person",
"chair_or_seat",
"instrument"
]
} | [
{
"area": 58800,
"bbox": [
1149,
0,
131,
606
],
"category_id": 0,
"id": 18639,
"image_id": "216_ln63SOGqztc_00001968",
"iscrowd": 0,
"segmentation": {
"counts": "`lWi03]f07I<D7I<D6J=C6J=C6J=C6J=C6J<D7I<D7I<D6J=C6J=C6J<D7I=C6J<D6J=C7I<D6J=C6J=C6J<D7I=C6J<... | [
{
"id": 1,
"name": "object"
}
] |
train | 2172_p7kSrenOmNU_00003033 | 2172_p7kSrenOmNU_00003033.jpg | {
"data_source": "VIPSeg",
"file_name": "2172_p7kSrenOmNU_00003033.jpg",
"height": 720,
"id": "2172_p7kSrenOmNU_00003033",
"width": 1280
} | {
"caption": "A person wearing a light pink headscarf and a striped skirt is playing a keyboard instrument with their right arm extended over the keys, while sitting on a chair with a white chair visible behind them on the dark floor.",
"caption_ann": "A <1:person wearing a light pink headscarf and a striped skirt> is playing a <4:keyboard instrument> with their right arm extended over the keys, while sitting on a <2:chair> with a <3:white chair> visible behind them on the <0:dark floor>.",
"id": 1724,
"image_id": "2172_p7kSrenOmNU_00003033",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "person wearing a light pink headscarf and a striped skirt"
},
{
"mask_ids": [
4
],
"txt_desc": "keyboard instrument"
},
{
"mask_ids": [
2
],
"txt_desc": "chair"
},
{
"mask_ids": [
3
],
"txt_desc": "white chair"
},
{
"mask_ids": [
0
],
"txt_desc": "dark floor"
}
],
"labels": [
"floor",
"person",
"chair_or_seat",
"chair_or_seat",
"instrument"
]
} | [
{
"area": 74959,
"bbox": [
537,
1,
593,
494
],
"category_id": 13,
"id": 18647,
"image_id": "2172_p7kSrenOmNU_00003033",
"iscrowd": 0,
"segmentation": {
"counts": "bbi;9Wf0c0]O00000000O1000_^OAQ=?oB5]<KcCi0i;WOWDT1^;lNbDT1^;lNbDT1^;lNbDT1^;lNbDT1^;lNbDT1^... | [
{
"id": 1,
"name": "object"
}
] |
train | 2173_0gf6rtImx0w_00000861 | 2173_0gf6rtImx0w_00000861.jpg | {
"data_source": "VIPSeg",
"file_name": "2173_0gf6rtImx0w_00000861.jpg",
"height": 720,
"id": "2173_0gf6rtImx0w_00000861",
"width": 1280
} | {
"caption": "A person in a red sari and black shawl is demonstrating a silver refrigerator with a transparent door panel, while standing in front of a light-colored wall with a round clock hanging on it, and to their left are a another silver refrigerator and a other refrigerator in the background, with a dark table or desk and a small dark trash can near the person, under a white ceiling.",
"caption_ann": "A <3:person in a red sari and black shawl> is demonstrating a <6:silver refrigerator with a transparent door panel>, while standing in front of a <0:light-colored wall> with a <8:round clock> hanging on it, and to their left are a <7: another silver refrigerator> and a <5:other refrigerator> in the background, with a <4:dark table or desk> and a <2:small dark trash can> near the person, under a <1:white ceiling>.",
"id": 1725,
"image_id": "2173_0gf6rtImx0w_00000861",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "person in a red sari and black shawl"
},
{
"mask_ids": [
6
],
"txt_desc": "silver refrigerator with a transparent door panel"
},
{
"mask_ids": [
0
],
"txt_desc": "light-colored wall"
},
{
"mask_ids": [
8
],
"txt_desc": "round clock"
},
{
"mask_ids": [
7
],
"txt_desc": "another silver refrigerator"
},
{
"mask_ids": [
5
],
"txt_desc": "other refrigerator"
},
{
"mask_ids": [
4
],
"txt_desc": "dark table or desk"
},
{
"mask_ids": [
2
],
"txt_desc": "small dark trash can"
},
{
"mask_ids": [
1
],
"txt_desc": "white ceiling"
}
],
"labels": [
"wall",
"ceiling",
"trash_can",
"person",
"table_or_desk",
"refrigerator",
"refrigerator",
"refrigerator",
"clock"
]
} | [
{
"area": 41145,
"bbox": [
843,
0,
437,
363
],
"category_id": 0,
"id": 18652,
"image_id": "2173_0gf6rtImx0w_00000861",
"iscrowd": 0,
"segmentation": {
"counts": "`g`b01_f0001O00001O001O00001O00001O00001O0000001O00001O0000001O00001O00001O00001O001O001O001... | [
{
"id": 1,
"name": "object"
}
] |
train | 2174_9IxPX6I_Hx8_00003120 | 2174_9IxPX6I_Hx8_00003120.jpg | {
"data_source": "VIPSeg",
"file_name": "2174_9IxPX6I_Hx8_00003120.jpg",
"height": 720,
"id": "2174_9IxPX6I_Hx8_00003120",
"width": 1280
} | {
"caption": "A person in a blue and white plaid shirt is standing next to an open black refrigerator with its freezer door also open, while looking towards a window that offers a view of buildings outside, all within a room with a white wall to the right, a yellow door with a gold knob, a light-colored floor, and a brick wall partially visible.",
"caption_ann": "A <4:person in a blue and white plaid shirt> is standing next to an open <5:black refrigerator> with its freezer door also open, while looking towards a <3:window> that offers a view of buildings outside, all within a room with a <0:white wall> to the right, a <2:yellow door> with a gold knob, a <1:light-colored floor>, and a brick wall partially visible.",
"id": 1726,
"image_id": "2174_9IxPX6I_Hx8_00003120",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "person in a blue and white plaid shirt"
},
{
"mask_ids": [
5
],
"txt_desc": "black refrigerator"
},
{
"mask_ids": [
3
],
"txt_desc": "window"
},
{
"mask_ids": [
0
],
"txt_desc": "white wall"
},
{
"mask_ids": [
2
],
"txt_desc": "yellow door"
},
{
"mask_ids": [
1
],
"txt_desc": "light-colored floor"
}
],
"labels": [
"wall",
"floor",
"door",
"window",
"person",
"refrigerator"
]
} | [
{
"area": 220679,
"bbox": [
0,
0,
1215,
720
],
"category_id": 0,
"id": 18661,
"image_id": "2174_9IxPX6I_Hx8_00003120",
"iscrowd": 0,
"segmentation": {
"counts": "0^2Y5[1l0`3TO_Lm0a3TO]Ln0b3RO]LP1b3QO\\LP1c3QO\\LR1b3nN]LU1a3mN[LW1c3jN[LY1b3iN[L\\1a3eN_L\\... | [
{
"id": 1,
"name": "object"
}
] |
train | 2175_FV3TSGdAaqw_00005511 | 2175_FV3TSGdAaqw_00005511.jpg | {
"data_source": "VIPSeg",
"file_name": "2175_FV3TSGdAaqw_00005511.jpg",
"height": 720,
"id": "2175_FV3TSGdAaqw_00005511",
"width": 1280
} | {
"caption": "A person wearing a blue baseball cap and a dusty blue shirt is opening a silver refrigerator in a kitchen with light-colored wall tiles and white cupboards or storage racks. On the left side a countertop or table holds various items including a two dark wine bottles, a champagne bottle, a white bottle, an unidentified bottle, and two bags or packages, as well as tissue. To the right two towels hang on the kitchen cabinets. ",
"caption_ann": "A <3:person wearing a blue baseball cap and a dusty blue shirt> is opening a <13:silver refrigerator> in a kitchen with <0:light-colored wall tiles> and <1:white cupboards or storage racks>. On the left side a <12:countertop or table> holds various items including a <7,8:two dark wine bottles>, a <9:champagne bottle>, a <10:white bottle>, an <11:unidentified bottle>, and two <4,5:bags or packages>, as well as <2:tissue>. To the right <6:two towels> hang on the <1:kitchen cabinets>. ",
"id": 1727,
"image_id": "2175_FV3TSGdAaqw_00005511",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "person wearing a blue baseball cap and a dusty blue shirt"
},
{
"mask_ids": [
13
],
"txt_desc": "silver refrigerator"
},
{
"mask_ids": [
0
],
"txt_desc": "light-colored wall tiles"
},
{
"mask_ids": [
1
],
"txt_desc": "white cupboards or storage racks"
},
{
"mask_ids": [
12
],
"txt_desc": "countertop or table"
},
{
"mask_ids": [
7,
8
],
"txt_desc": "two dark wine bottles"
},
{
"mask_ids": [
9
],
"txt_desc": "champagne bottle"
},
{
"mask_ids": [
10
],
"txt_desc": "white bottle"
},
{
"mask_ids": [
11
],
"txt_desc": "unidentified bottle"
},
{
"mask_ids": [
4,
5
],
"txt_desc": "bags or packages"
},
{
"mask_ids": [
2
],
"txt_desc": "tissue"
},
{
"mask_ids": [
6
],
"txt_desc": "two towels"
},
{
"mask_ids": [
1
],
"txt_desc": "kitchen cabinets"
}
],
"labels": [
"wall",
"cupboard_or_showcase_or_storage_rack",
"tissue",
"person",
"bag_or_package",
"bag_or_package",
"bag_or_package",
"bottle_or_cup",
"bottle_or_cup",
"bottle_or_cup",
"bottle_or_cup",
"bottle_or_cup",
"table_or_desk",
"refrigerator"
]
} | [
{
"area": 50227,
"bbox": [
0,
299,
1280,
421
],
"category_id": 0,
"id": 18667,
"image_id": "2175_FV3TSGdAaqw_00005511",
"iscrowd": 0,
"segmentation": {
"counts": "e9i3gb0>B;Ej0VO;Ea0_O2N2N00000O100000O100001O000000WMn@TNR?h1WASNi>h1iAkMW>S2mAjMS>S2VBhMj=... | [
{
"id": 1,
"name": "object"
}
] |
train | 2180_jZEPJa15A10_00000384 | 2180_jZEPJa15A10_00000384.jpg | {
"data_source": "VIPSeg",
"file_name": "2180_jZEPJa15A10_00000384.jpg",
"height": 720,
"id": "2180_jZEPJa15A10_00000384",
"width": 1280
} | {
"caption": "A person with a beard, wearing a black t-shirt, is pointing at the dispenser of a silver refrigerator, while standing in a room with a green wall and a white ceiling, where a red flower arrangement and green plants are visible atop a dark cupboard or storage rack to the left.",
"caption_ann": "A <5:person with a beard, wearing a black t-shirt>, is pointing at the dispenser of a <6:silver refrigerator>, while standing in a room with a <0:green wall> and a <1:white ceiling>, where a <2:red flower arrangement> and <3:green plants> are visible atop a <4:dark cupboard or storage rack> to the left.",
"id": 1728,
"image_id": "2180_jZEPJa15A10_00000384",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "person with a beard, wearing a black t-shirt"
},
{
"mask_ids": [
6
],
"txt_desc": "silver refrigerator"
},
{
"mask_ids": [
0
],
"txt_desc": "green wall"
},
{
"mask_ids": [
1
],
"txt_desc": "white ceiling"
},
{
"mask_ids": [
2
],
"txt_desc": "red flower arrangement"
},
{
"mask_ids": [
3
],
"txt_desc": "green plants"
},
{
"mask_ids": [
4
],
"txt_desc": "dark cupboard or storage rack"
}
],
"labels": [
"wall",
"ceiling",
"flower",
"other_plant",
"cupboard_or_showcase_or_storage_rack",
"person",
"refrigerator"
]
} | [
{
"area": 110258,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18681,
"image_id": "2180_jZEPJa15A10_00000384",
"iscrowd": 0,
"segmentation": {
"counts": "[4Ub0[4O1ZOf0[Od0H9I8K4L5J6K4L4I7B>K5J6B>H7I8G8K5K6L3J7H7I8I6L5I7F:E;G9G:@?K6L3L4K5G9H9I6I7E... | [
{
"id": 1,
"name": "object"
}
] |
train | 2181_p5dXL3TsyVw_00000379 | 2181_p5dXL3TsyVw_00000379.jpg | {
"data_source": "VIPSeg",
"file_name": "2181_p5dXL3TsyVw_00000379.jpg",
"height": 720,
"id": "2181_p5dXL3TsyVw_00000379",
"width": 1280
} | {
"caption": "A person wearing a dark grey zippered jacket and a black polo shirt is presenting a large silver refrigerator with a price tag of $2497.00, while standing in a store with a white wall in the background, where other refrigerators are lined up, and a fluorescent lamp provides illumination over a light-colored floor.",
"caption_ann": "A <3:person wearing a dark grey zippered jacket and a black polo shirt> is presenting a <4:large silver refrigerator> with a price tag of $2497.00, while standing in a store with a <0:white wall> in the background, where other <5,6,7,8:refrigerators> are lined up, and a <2:fluorescent lamp> provides illumination over a <1:light-colored floor>.",
"id": 1729,
"image_id": "2181_p5dXL3TsyVw_00000379",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "person wearing a dark grey zippered jacket and a black polo shirt"
},
{
"mask_ids": [
4
],
"txt_desc": "large silver refrigerator"
},
{
"mask_ids": [
0
],
"txt_desc": "white wall"
},
{
"mask_ids": [
5,
6,
7,
8
],
"txt_desc": "refrigerators"
},
{
"mask_ids": [
2
],
"txt_desc": "fluorescent lamp"
},
{
"mask_ids": [
1
],
"txt_desc": "light-colored floor"
}
],
"labels": [
"wall",
"floor",
"lamp",
"person",
"refrigerator",
"refrigerator",
"refrigerator",
"refrigerator",
"refrigerator"
]
} | [
{
"area": 52371,
"bbox": [
160,
0,
1120,
181
],
"category_id": 0,
"id": 18688,
"image_id": "2181_p5dXL3TsyVw_00000379",
"iscrowd": 0,
"segmentation": {
"counts": "P``31_f000000000000000000000000000001O0000001O0000000000001O00001O00000000000000000000001O0... | [
{
"id": 1,
"name": "object"
}
] |
train | 2183_9Jq-7HnWa_0_00003288 | 2183_9Jq-7HnWa_0_00003288.jpg | {
"data_source": "VIPSeg",
"file_name": "2183_9Jq-7HnWa_0_00003288.jpg",
"height": 720,
"id": "2183_9Jq-7HnWa_0_00003288",
"width": 1280
} | {
"caption": "A person with short grey hair and a beige sweater with an American flag design is standing in a kitchen with a patterned backsplash wall and white upper cabinets, preparing a raw turkey in a white roasting pan, with a roaster to the right, a microwave oven above the stove, a dark countertop with a carrot and a silver bowl, and a blue bottle and a green bottle near a green tool, while a floral curtain hangs in the background, all beside a wooden table.",
"caption_ann": "A <5:person with short grey hair and a beige sweater with an American flag design> is standing in a kitchen with <0:a patterned backsplash wall> and <1:white upper cabinets>, preparing a <7:raw turkey> in a <6:white roasting pan>, with a <13:roaster> to the right, a <14:microwave oven> above the stove, a <12:dark countertop> with a <4:carrot> and a <8:silver bowl>, and a <9:blue bottle> and a <10:green bottle> near a <3:green tool>, while a <2:floral curtain> hangs in the background, all beside a <11:wooden table>.",
"id": 1730,
"image_id": "2183_9Jq-7HnWa_0_00003288",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "person with short grey hair and a beige sweater with an American flag design"
},
{
"mask_ids": [
0
],
"txt_desc": "a patterned backsplash wall"
},
{
"mask_ids": [
1
],
"txt_desc": "white upper cabinets"
},
{
"mask_ids": [
7
],
"txt_desc": "raw turkey"
},
{
"mask_ids": [
6
],
"txt_desc": "white roasting pan"
},
{
"mask_ids": [
13
],
"txt_desc": "roaster"
},
{
"mask_ids": [
14
],
"txt_desc": "microwave oven"
},
{
"mask_ids": [
12
],
"txt_desc": "dark countertop"
},
{
"mask_ids": [
4
],
"txt_desc": "carrot"
},
{
"mask_ids": [
8
],
"txt_desc": "silver bowl"
},
{
"mask_ids": [
9
],
"txt_desc": "blue bottle"
},
{
"mask_ids": [
10
],
"txt_desc": "green bottle"
},
{
"mask_ids": [
3
],
"txt_desc": "green tool"
},
{
"mask_ids": [
2
],
"txt_desc": "floral curtain"
},
{
"mask_ids": [
11
],
"txt_desc": "wooden table"
}
],
"labels": [
"wall",
"cupboard_or_showcase_or_storage_rack",
"curtain",
"tool",
"food",
"person",
"tub_or_bowl_or_pot",
"tub_or_bowl_or_pot",
"tub_or_bowl_or_pot",
"bottle_or_cup",
"bottle_or_cup",
"table_or_desk",
"table_or_desk",
"roaster",
"Microwave_oven"
]
} | [
{
"area": 116779,
"bbox": [
0,
266,
1280,
275
],
"category_id": 0,
"id": 18697,
"image_id": "2183_9Jq-7HnWa_0_00003288",
"iscrowd": 0,
"segmentation": {
"counts": "W:h0ge0S4nK1O00000O100000000O10000006J=C<D=C=C=C=C<E<C7IO100000000O10000000000O100000000O2... | [
{
"id": 1,
"name": "object"
}
] |
train | 2186_C8NTYa5Cf2I_00001248 | 2186_C8NTYa5Cf2I_00001248.jpg | {
"data_source": "VIPSeg",
"file_name": "2186_C8NTYa5Cf2I_00001248.jpg",
"height": 720,
"id": "2186_C8NTYa5Cf2I_00001248",
"width": 1280
} | {
"caption": "A white shelf unit with a round green plant on its top shelf and folded white towels on its two lower shelves stands next to a white bathtub with tiled walls in the background, all resting on a tiled floor, with four clear jars with white lids on the very top of the shelf.",
"caption_ann": "A <3:white shelf unit> with a <2:round green plant> on its top shelf and <5:folded white towels> on its two lower shelves stands next to a <4:white bathtub> with <0:tiled walls> in the background, all resting on a <1:tiled floor>, with <6,7,8,9:four clear jars with white lids> on the very top of the shelf.",
"id": 1731,
"image_id": "2186_C8NTYa5Cf2I_00001248",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "white shelf unit"
},
{
"mask_ids": [
2
],
"txt_desc": "round green plant"
},
{
"mask_ids": [
5
],
"txt_desc": "folded white towels"
},
{
"mask_ids": [
4
],
"txt_desc": "white bathtub"
},
{
"mask_ids": [
0
],
"txt_desc": "tiled walls"
},
{
"mask_ids": [
1
],
"txt_desc": "tiled floor"
},
{
"mask_ids": [
6,
7,
8,
9
],
"txt_desc": "four clear jars with white lids"
}
],
"labels": [
"wall",
"floor",
"other_plant",
"shelf",
"bathtub",
"textiles",
"bottle_or_cup",
"bottle_or_cup",
"bottle_or_cup",
"bottle_or_cup"
]
} | [
{
"area": 544780,
"bbox": [
0,
0,
1280,
693
],
"category_id": 0,
"id": 18712,
"image_id": "2186_C8NTYa5Cf2I_00001248",
"iscrowd": 0,
"segmentation": {
"counts": "0i:g;N20000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 2187_2_OkSoW2tkA_00002268 | 2187_2_OkSoW2tkA_00002268.jpg | {
"data_source": "VIPSeg",
"file_name": "2187_2_OkSoW2tkA_00002268.jpg",
"height": 720,
"id": "2187_2_OkSoW2tkA_00002268",
"width": 1280
} | {
"caption": "A traffic light displaying a red signal hangs from a tall grey pole with numerous wires extending across the clear blue sky, while a metal fence is visible on the left, and a white bridge is situated to the right, all above a grassy field and in front of a distant tree-covered mountain.",
"caption_ann": "A <5:traffic light displaying a red signal> hangs from a <1:tall grey pole> with numerous wires extending across the <4:clear blue sky>, while a <0:metal fence> is visible on the left, and a <3:white bridge> is situated to the right, all above a <2:grassy field> and in front of a distant <6:tree-covered mountain>.",
"id": 1732,
"image_id": "2187_2_OkSoW2tkA_00002268",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "traffic light displaying a red signal"
},
{
"mask_ids": [
1
],
"txt_desc": "tall grey pole"
},
{
"mask_ids": [
4
],
"txt_desc": "clear blue sky"
},
{
"mask_ids": [
0
],
"txt_desc": "metal fence"
},
{
"mask_ids": [
3
],
"txt_desc": "white bridge"
},
{
"mask_ids": [
2
],
"txt_desc": "grassy field"
},
{
"mask_ids": [
6
],
"txt_desc": "tree-covered mountain"
}
],
"labels": [
"handrail_or_fence",
"pole",
"grass",
"bridge",
"sky",
"traffic_light",
"tree"
]
} | [
{
"area": 45138,
"bbox": [
0,
470,
638,
153
],
"category_id": 7,
"id": 18722,
"image_id": "2187_2_OkSoW2tkA_00002268",
"iscrowd": 0,
"segmentation": {
"counts": "f>i4ha0N100000001O000000001O000000001O000000001O000000001O0000001O000000001O000000001O000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 2189_8jpjltOLbNs_00000618 | 2189_8jpjltOLbNs_00000618.jpg | {
"data_source": "VIPSeg",
"file_name": "2189_8jpjltOLbNs_00000618.jpg",
"height": 720,
"id": "2189_8jpjltOLbNs_00000618",
"width": 1280
} | {
"caption": "A large, ornate white building with two domed red roofs serves as a grand entrance. A traffic light displaying a green signal hangs in front of the building above a crosswalk, as an green jeep, a white car, and a silver car drive along the city road. Two guards are standing near the entrance, and several flower pots decorate the sidewalk with plants, while poles and distant trees frame the scene under a bright sky. A white car moves in the foreground on the left. The \"GoodNews24\" logo suggests this is a TV news broadcast.",
"caption_ann": "A <4:large, ornate white building with two domed red roofs> serves as a grand entrance. A <6:traffic light displaying a green signal> hangs in front of the <4:building> above a <3:crosswalk>, as an <11:green jeep>, a <10:white car>, and a <9:silver car> drive along the <2:city road>. Two <13,14:guards> are standing near the entrance, and several <15,16,17,18,19:flower pots> decorate the <1:sidewalk> with <8:plants>, while <0:poles> and <7:distant trees> frame the scene under a <5:bright sky>. A <12:white car> moves in the foreground on the left. The \"GoodNews24\" logo suggests this is a TV news broadcast.",
"id": 1733,
"image_id": "2189_8jpjltOLbNs_00000618",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "large, ornate white building with two domed red roofs"
},
{
"mask_ids": [
6
],
"txt_desc": "traffic light displaying a green signal"
},
{
"mask_ids": [
4
],
"txt_desc": "building"
},
{
"mask_ids": [
3
],
"txt_desc": "crosswalk"
},
{
"mask_ids": [
11
],
"txt_desc": "green jeep"
},
{
"mask_ids": [
10
],
"txt_desc": "white car"
},
{
"mask_ids": [
9
],
"txt_desc": "silver car"
},
{
"mask_ids": [
2
],
"txt_desc": "city road"
},
{
"mask_ids": [
13,
14
],
"txt_desc": "guards"
},
{
"mask_ids": [
15,
16,
17,
18,
19
],
"txt_desc": "flower pots"
},
{
"mask_ids": [
1
],
"txt_desc": "sidewalk"
},
{
"mask_ids": [
8
],
"txt_desc": "plants"
},
{
"mask_ids": [
0
],
"txt_desc": "poles"
},
{
"mask_ids": [
7
],
"txt_desc": "distant trees"
},
{
"mask_ids": [
5
],
"txt_desc": "bright sky"
},
{
"mask_ids": [
12
],
"txt_desc": "white car"
}
],
"labels": [
"pole",
"ground",
"road",
"crosswalk",
"other_construction",
"sky",
"traffic_light",
"tree",
"other_plant",
"car",
"car",
"car",
"car",
"person",
"person",
"flower_pot_or_vase",
"flower_pot_or_vase",
"flower_pot_or_vase",
"flower_pot_or_vase",
"flower_pot_or_vase"
]
} | [
{
"area": 23015,
"bbox": [
310,
0,
808,
636
],
"category_id": 12,
"id": 18729,
"image_id": "2189_8jpjltOLbNs_00000618",
"iscrowd": 0,
"segmentation": {
"counts": "dYj6Y4m`0Y5\\Kf0eN\\1XOf0fN\\1YOe0fN\\1YOf0eN[1D;M3OiLkKUDR4l;SLQAk0k1Q3T=nMXBo1S9_MkHd1PMn... | [
{
"id": 1,
"name": "object"
}
] |
train | 2190_3kMlEOxf2qE_00000048 | 2190_3kMlEOxf2qE_00000048.jpg | {
"data_source": "VIPSeg",
"file_name": "2190_3kMlEOxf2qE_00000048.jpg",
"height": 720,
"id": "2190_3kMlEOxf2qE_00000048",
"width": 1280
} | {
"caption": "A person wearing a blue shirt and dark track pants with red stripes is working on a dark tire with a series of other tires stacked behind them on the left, while a metal shelf with various machine parts and a white wall are visible to the right, all above a dark floor.",
"caption_ann": "A <5:person wearing a blue shirt and dark track pants with red stripes> is working on a <2:dark tire> with a series of <2:other tires> stacked behind them on the left, while a <3:metal shelf> with various <4:machine parts> and a <0:white wall> are visible to the right, all above a <1:dark floor>.",
"id": 1734,
"image_id": "2190_3kMlEOxf2qE_00000048",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "person wearing a blue shirt and dark track pants with red stripes"
},
{
"mask_ids": [
2
],
"txt_desc": "dark tire"
},
{
"mask_ids": [
2
],
"txt_desc": "other tires"
},
{
"mask_ids": [
3
],
"txt_desc": "metal shelf"
},
{
"mask_ids": [
4
],
"txt_desc": "machine parts"
},
{
"mask_ids": [
0
],
"txt_desc": "white wall"
},
{
"mask_ids": [
1
],
"txt_desc": "dark floor"
}
],
"labels": [
"wall",
"floor",
"tyre",
"shelf",
"other_machine",
"person"
]
} | [
{
"area": 39211,
"bbox": [
405,
0,
875,
570
],
"category_id": 0,
"id": 18749,
"image_id": "2190_3kMlEOxf2qE_00000048",
"iscrowd": 0,
"segmentation": {
"counts": "`hl82^f0000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 2191_7mrySM_bNVk_00000348 | 2191_7mrySM_bNVk_00000348.jpg | {
"data_source": "VIPSeg",
"file_name": "2191_7mrySM_bNVk_00000348.jpg",
"height": 720,
"id": "2191_7mrySM_bNVk_00000348",
"width": 1280
} | {
"caption": "A person wearing a blue shirt and black pants is crouching on the red-brown tiled floor, using a red and black tool to work on a dark tire. Other tools, including a yellow and black hand saw, are scattered around the person on the floor, with a white wall visible in the background.",
"caption_ann": "A <4:person wearing a blue shirt and black pants> is crouching on the <1:red-brown tiled floor>, using a <3:red and black tool> to work on a <2:dark tire>. Other <3:tools>, including a <3:yellow and black hand saw>, are scattered around the <4:person> on the <1:floor>, with a <0:white wall> visible in the background.",
"id": 1735,
"image_id": "2191_7mrySM_bNVk_00000348",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "person wearing a blue shirt and black pants"
},
{
"mask_ids": [
1
],
"txt_desc": "red-brown tiled floor"
},
{
"mask_ids": [
3
],
"txt_desc": "red and black tool"
},
{
"mask_ids": [
2
],
"txt_desc": "dark tire"
},
{
"mask_ids": [
3
],
"txt_desc": "tools"
},
{
"mask_ids": [
3
],
"txt_desc": "yellow and black hand saw"
},
{
"mask_ids": [
4
],
"txt_desc": "person"
},
{
"mask_ids": [
1
],
"txt_desc": "floor"
},
{
"mask_ids": [
0
],
"txt_desc": "white wall"
}
],
"labels": [
"wall",
"floor",
"tyre",
"tool",
"person"
]
} | [
{
"area": 8728,
"bbox": [
1185,
0,
95,
182
],
"category_id": 0,
"id": 18755,
"image_id": "2191_7mrySM_bNVk_00000348",
"iscrowd": 0,
"segmentation": {
"counts": "`VQj01_f02N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N2N2N2N2N2N2N2N2N2N1O2N2N2N... | [
{
"id": 1,
"name": "object"
}
] |
train | 2193_0lpFur6cd_M_00003168 | 2193_0lpFur6cd_M_00003168.jpg | {
"data_source": "VIPSeg",
"file_name": "2193_0lpFur6cd_M_00003168.jpg",
"height": 720,
"id": "2193_0lpFur6cd_M_00003168",
"width": 1280
} | {
"caption": "A person's hands are arranging a large bouquet of vibrant pink flowers and green foliage into a clear glass vase on a round wooden table, with another pink flower in a smaller clear glass vase to the right against a light-colored wall.",
"caption_ann": "A <2:person's hands> are arranging a <1:large bouquet of vibrant pink flowers> and <1:green foliage> into a <5:clear glass vase> on a <3:round wooden table>, with another <1:pink flower> in a <4:smaller clear glass vase> to the right against a <0:light-colored wall>.",
"id": 1736,
"image_id": "2193_0lpFur6cd_M_00003168",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "person's hands"
},
{
"mask_ids": [
1
],
"txt_desc": "large bouquet of vibrant pink flowers"
},
{
"mask_ids": [
1
],
"txt_desc": "green foliage"
},
{
"mask_ids": [
5
],
"txt_desc": "clear glass vase"
},
{
"mask_ids": [
3
],
"txt_desc": "round wooden table"
},
{
"mask_ids": [
1
],
"txt_desc": "pink flower"
},
{
"mask_ids": [
4
],
"txt_desc": "smaller clear glass vase"
},
{
"mask_ids": [
0
],
"txt_desc": "light-colored wall"
}
],
"labels": [
"wall",
"flower",
"person",
"table_or_desk",
"flower_pot_or_vase",
"flower_pot_or_vase"
]
} | [
{
"area": 290264,
"bbox": [
28,
0,
1252,
391
],
"category_id": 0,
"id": 18760,
"image_id": "2193_0lpFur6cd_M_00003168",
"iscrowd": 0,
"segmentation": {
"counts": "inc0c0bc0[2kNU1eM[2kNU1WNi100000001O00000000000000000000000000000000001O0000000000000000001... | [
{
"id": 1,
"name": "object"
}
] |
train | 219_-70GtKnhvos_00004713 | 219_-70GtKnhvos_00004713.jpg | {
"data_source": "VIPSeg",
"file_name": "219_-70GtKnhvos_00004713.jpg",
"height": 720,
"id": "219_-70GtKnhvos_00004713",
"width": 1280
} | {
"caption": "A man in a black hoodie and a young boy in a blue and red soccer jersey sit at a grey table in front of a light coloured wall, with two silver cloches on white plates in front of them, flanked by a dark blue chair and a dark grey sofa.",
"caption_ann": "A <2:man in a black hoodie> and a <1:young boy in a blue and red soccer jersey> sit at a <7:grey table> in front of a <0:light coloured wall>, with <5,6:two silver cloches> on <3,4:white plates> in front of them, flanked by a <8:dark blue chair> and a <9:dark grey sofa>.",
"id": 1737,
"image_id": "219_-70GtKnhvos_00004713",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "man in a black hoodie"
},
{
"mask_ids": [
1
],
"txt_desc": "young boy in a blue and red soccer jersey"
},
{
"mask_ids": [
7
],
"txt_desc": "grey table"
},
{
"mask_ids": [
0
],
"txt_desc": "light coloured wall"
},
{
"mask_ids": [
5,
6
],
"txt_desc": "two silver cloches"
},
{
"mask_ids": [
3,
4
],
"txt_desc": "white plates"
},
{
"mask_ids": [
8
],
"txt_desc": "dark blue chair"
},
{
"mask_ids": [
9
],
"txt_desc": "dark grey sofa"
}
],
"labels": [
"wall",
"person",
"person",
"plate",
"plate",
"tub_or_bowl_or_pot",
"tub_or_bowl_or_pot",
"table_or_desk",
"chair_or_seat",
"sofa"
]
} | [
{
"area": 323141,
"bbox": [
0,
0,
1280,
489
],
"category_id": 0,
"id": 18766,
"image_id": "219_-70GtKnhvos_00004713",
"iscrowd": 0,
"segmentation": {
"counts": "0Y7W?1O00000000000000000000000000000000001O00000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 21_WNfqB3toDh8_00000436 | 21_WNfqB3toDh8_00000436.jpg | {
"data_source": "VIPSeg",
"file_name": "21_WNfqB3toDh8_00000436.jpg",
"height": 720,
"id": "21_WNfqB3toDh8_00000436",
"width": 1280
} | {
"caption": "On a green grass field, a group of ten women are engaged in a soccer game. A player in a blue shirt dribbles a soccer ball, while other nine players, mostly wearing bright green jerseys, position themselves around the field. In the background, there is a running track with bleacher-style stairs, a low fence, and a tall pole. Beyond the field, numerous trees stand under a hazy sky.",
"caption_ann": "On a <3:green grass field>, a <6,7,8,9,10,11,12,13,14,15:group of ten women> are engaged in a soccer game. A <11:player in a blue shirt> dribbles a <16:soccer ball>, while other <6,7,8,9,10,12,13,14,15:nine players>, mostly wearing bright green jerseys, position themselves around the <3:field>. In the background, there is a <0:running track with bleacher-style stairs>, a low <1:fence>, and a <2:tall pole>. Beyond the field, <5:numerous trees> stand under a <4:hazy sky>.",
"id": 1738,
"image_id": "21_WNfqB3toDh8_00000436",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "green grass field"
},
{
"mask_ids": [
6,
7,
8,
9,
10,
11,
12,
13,
14,
15
],
"txt_desc": "group of ten women"
},
{
"mask_ids": [
11
],
"txt_desc": "player in a blue shirt"
},
{
"mask_ids": [
16
],
"txt_desc": "soccer ball"
},
{
"mask_ids": [
6,
7,
8,
9,
10,
12,
13,
14,
15
],
"txt_desc": "nine players"
},
{
"mask_ids": [
3
],
"txt_desc": "field"
},
{
"mask_ids": [
0
],
"txt_desc": "running track with bleacher-style stairs"
},
{
"mask_ids": [
1
],
"txt_desc": "fence"
},
{
"mask_ids": [
2
],
"txt_desc": "tall pole"
},
{
"mask_ids": [
5
],
"txt_desc": "numerous trees"
},
{
"mask_ids": [
4
],
"txt_desc": "hazy sky"
}
],
"labels": [
"stair",
"handrail_or_fence",
"pole",
"grass",
"sky",
"tree",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"ball"
]
} | [
{
"area": 45510,
"bbox": [
0,
208,
1280,
126
],
"category_id": 3,
"id": 18776,
"image_id": "21_WNfqB3toDh8_00000436",
"iscrowd": 0,
"segmentation": {
"counts": "`6`1Pe000000000000000000000000001O000001O00N2N2O1N2N2O1O1O2O0O1O1O1QOgZO>cg>EbnA4L3N0O1O001O0... | [
{
"id": 1,
"name": "object"
}
] |
train | 222_XAWNzd-gc3Q_00000123 | 222_XAWNzd-gc3Q_00000123.jpg | {
"data_source": "VIPSeg",
"file_name": "222_XAWNzd-gc3Q_00000123.jpg",
"height": 720,
"id": "222_XAWNzd-gc3Q_00000123",
"width": 1280
} | {
"caption": "In a room decorated with seven graphic card boxes for various computer hardware components displayed on a shelf against a plain wall, a man in a blue shirt and black hat stands at a white table to present an empty black CPU cabinet.",
"caption_ann": "In a room decorated with <2,3,4,5,6,7,8:seven graphic card boxes for various computer hardware components> displayed on a shelf against a <0:plain wall>, a <1:man in a blue shirt and black hat> stands at a <9:white table> to present an empty <10:black CPU cabinet>.",
"id": 1739,
"image_id": "222_XAWNzd-gc3Q_00000123",
"label_matched": [
{
"mask_ids": [
2,
3,
4,
5,
6,
7,
8
],
"txt_desc": "seven graphic card boxes for various computer hardware components"
},
{
"mask_ids": [
0
],
"txt_desc": "plain wall"
},
{
"mask_ids": [
1
],
"txt_desc": "man in a blue shirt and black hat"
},
{
"mask_ids": [
9
],
"txt_desc": "white table"
},
{
"mask_ids": [
10
],
"txt_desc": "black CPU cabinet"
}
],
"labels": [
"wall",
"person",
"box",
"box",
"box",
"box",
"box",
"box",
"box",
"table_or_desk",
"computer"
]
} | [
{
"area": 490740,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18793,
"image_id": "222_XAWNzd-gc3Q_00000123",
"iscrowd": 0,
"segmentation": {
"counts": "U3[c0U300000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 223_X2LLju41exc_00002027 | 223_X2LLju41exc_00002027.jpg | {
"data_source": "VIPSeg",
"file_name": "223_X2LLju41exc_00002027.jpg",
"height": 720,
"id": "223_X2LLju41exc_00002027",
"width": 1280
} | {
"caption": "Inside a dimly lit space defined by rough wooden walls and a dirt floor, a group of people gathers around large wired cages holding several green parakeets. The scene is viewed from the close perspective of a guy in a pink patterned t-shirt, with other onlookers including a person in black and white patterned shirt, a guy with a backpack, and a man in a blue patterned t-shirt standing nearby.",
"caption_ann": "Inside a dimly lit space defined by <0:rough wooden walls> and a <1:dirt floor>, a group of people gathers around <2:large wired cages> holding <7,8:several green parakeets>. The scene is viewed from the close perspective of a <3:guy in a pink patterned t-shirt>, with other onlookers including a <4:person in black and white patterned shirt>, a <5:guy with a backpack>, and a <6:man in a blue patterned t-shirt> standing nearby.",
"id": 1740,
"image_id": "223_X2LLju41exc_00002027",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "rough wooden walls"
},
{
"mask_ids": [
1
],
"txt_desc": "dirt floor"
},
{
"mask_ids": [
2
],
"txt_desc": "large wired cages"
},
{
"mask_ids": [
7,
8
],
"txt_desc": "several green parakeets"
},
{
"mask_ids": [
3
],
"txt_desc": "guy in a pink patterned t-shirt"
},
{
"mask_ids": [
4
],
"txt_desc": "person in black and white patterned shirt"
},
{
"mask_ids": [
5
],
"txt_desc": "guy with a backpack"
},
{
"mask_ids": [
6
],
"txt_desc": "man in a blue patterned t-shirt"
}
],
"labels": [
"wall",
"ground",
"cage",
"person",
"person",
"person",
"person",
"other_animal",
"other_animal"
]
} | [
{
"area": 178783,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18804,
"image_id": "223_X2LLju41exc_00002027",
"iscrowd": 0,
"segmentation": {
"counts": "0X;X;doe1YGRYZN2O1M3O1N2N4M8G6J8I2L5M1N5lLQL^CQ4`<SL\\CQ4_9jKhF8GP4]9jKjF<Dn3\\9iKoF<Bm3\\9iK... | [
{
"id": 1,
"name": "object"
}
] |
train | 223_YsIPo670N2A_00002718 | 223_YsIPo670N2A_00002718.jpg | {
"data_source": "VIPSeg",
"file_name": "223_YsIPo670N2A_00002718.jpg",
"height": 720,
"id": "223_YsIPo670N2A_00002718",
"width": 1280
} | {
"caption": "This image shows a backdrop of rocky, brush-covered mountains and a large sand berm under a light blue sky where a man in a black shirt and cap stands at an outdoor shooting range behind a light blue table. He is demonstrating tactical equipment laid out before him, including a camouflage vest with helmet, a military green pouch, a large backpack, and the tan tactical vest.",
"caption_ann": "This image shows a backdrop of <2:rocky, brush-covered mountains> and a <0:large sand berm> under a <1:light blue sky> where a <3:man in a black shirt and cap> stands at an outdoor shooting range behind a <8:light blue table>. He is demonstrating tactical equipment laid out before him, including a <4:camouflage vest with helmet>, a <6:military green pouch>, a <7:large backpack>, and the <5:tan tactical vest>.",
"id": 1741,
"image_id": "223_YsIPo670N2A_00002718",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "rocky, brush-covered mountains"
},
{
"mask_ids": [
0
],
"txt_desc": "large sand berm"
},
{
"mask_ids": [
1
],
"txt_desc": "light blue sky"
},
{
"mask_ids": [
3
],
"txt_desc": "man in a black shirt and cap"
},
{
"mask_ids": [
8
],
"txt_desc": "light blue table"
},
{
"mask_ids": [
4
],
"txt_desc": "camouflage vest with helmet"
},
{
"mask_ids": [
6
],
"txt_desc": "military green pouch"
},
{
"mask_ids": [
7
],
"txt_desc": "large backpack"
},
{
"mask_ids": [
5
],
"txt_desc": "tan tactical vest"
}
],
"labels": [
"sand",
"sky",
"mountain",
"person",
"bag_or_package",
"bag_or_package",
"bag_or_package",
"bag_or_package",
"table_or_desk"
]
} | [
{
"area": 406924,
"bbox": [
0,
129,
1280,
591
],
"category_id": 16,
"id": 18813,
"image_id": "223_YsIPo670N2A_00002718",
"iscrowd": 0,
"segmentation": {
"counts": "o6a?o600000000000000000000000000000000000O1000000000000000000O1000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 225_fwyvMqAwGA4_00000633 | 225_fwyvMqAwGA4_00000633.jpg | {
"data_source": "VIPSeg",
"file_name": "225_fwyvMqAwGA4_00000633.jpg",
"height": 720,
"id": "225_fwyvMqAwGA4_00000633",
"width": 1280
} | {
"caption": "A man wearing a black graphic t-shirt and light blue jeans stands behind a white table and holds up a white, square-shaped object for presentation. He is positioned in a room with a plain off-white wall, a white panelled door to his right, and a large grey barrel topped with a smaller blue container to his left.",
"caption_ann": "A <2:man wearing a black graphic t-shirt and light blue jeans> stands behind a <6:white table> and holds up a <3:white, square-shaped object> for presentation. He is positioned in a room with a <0:plain off-white wall>, a <1:white panelled door> to his right, and a <5:large grey barrel> topped with a <4:smaller blue container> to his left.",
"id": 1742,
"image_id": "225_fwyvMqAwGA4_00000633",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "man wearing a black graphic t-shirt and light blue jeans"
},
{
"mask_ids": [
6
],
"txt_desc": "white table"
},
{
"mask_ids": [
3
],
"txt_desc": "white, square-shaped object"
},
{
"mask_ids": [
0
],
"txt_desc": "plain off-white wall"
},
{
"mask_ids": [
1
],
"txt_desc": "white panelled door"
},
{
"mask_ids": [
5
],
"txt_desc": "large grey barrel"
},
{
"mask_ids": [
4
],
"txt_desc": "smaller blue container"
}
],
"labels": [
"wall",
"door",
"person",
"plate",
"barrel",
"barrel",
"table_or_desk"
]
} | [
{
"area": 614413,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18822,
"image_id": "225_fwyvMqAwGA4_00000633",
"iscrowd": 0,
"segmentation": {
"counts": "0jng24bgXM`0@`0@`0@?A`0@`0@?A`0@`0@`0@?A`0@`0@`0@?A`0@`0@?A`0@`0@`0UMbEdHm:l6TETI\\;[6eDeIk;j... | [
{
"id": 1,
"name": "object"
}
] |
train | 225_xAj8HBoTlrA_00002612 | 225_xAj8HBoTlrA_00002612.jpg | {
"data_source": "VIPSeg",
"file_name": "225_xAj8HBoTlrA_00002612.jpg",
"height": 720,
"id": "225_xAj8HBoTlrA_00002612",
"width": 1280
} | {
"caption": "A young boy wearing glasses and a black graphic t-shirt shares a quiet moment with a pet cockatiel perched on his arm. The interaction takes place in a room with a textured brown wall, where a large stack of birdcages stands in the background next to a wooden counter, and the legs of a small black table are visible at the bottom edge of the frame.",
"caption_ann": "A <2:young boy wearing glasses and a black graphic t-shirt> shares a quiet moment with a <3:pet cockatiel> perched on his arm. The interaction takes place in a room with a <0:textured brown wall>, where a <1:large stack of birdcages> stands in the background next to a <4:wooden counter>, and the <5:legs of a small black table> are visible at the bottom edge of the frame.",
"id": 1743,
"image_id": "225_xAj8HBoTlrA_00002612",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "young boy wearing glasses and a black graphic t-shirt"
},
{
"mask_ids": [
3
],
"txt_desc": "pet cockatiel"
},
{
"mask_ids": [
0
],
"txt_desc": "textured brown wall"
},
{
"mask_ids": [
1
],
"txt_desc": "large stack of birdcages"
},
{
"mask_ids": [
4
],
"txt_desc": "wooden counter"
},
{
"mask_ids": [
5
],
"txt_desc": "legs of a small black table"
}
],
"labels": [
"wall",
"cage",
"person",
"other_animal",
"table_or_desk",
"chair_or_seat"
]
} | [
{
"area": 383318,
"bbox": [
0,
2,
1280,
718
],
"category_id": 0,
"id": 18829,
"image_id": "225_xAj8HBoTlrA_00002612",
"iscrowd": 0,
"segmentation": {
"counts": "2d3g?`L[Ai1POg1e?`L[Ai1POg1e?`L\\Ah1oNh1e?`L\\Ah1oNh1e?`L\\Ah1oNh1e?`L]Ag1nNi1e?`L]Ag1nNi1e?`... | [
{
"id": 1,
"name": "object"
}
] |
train | 227_MU4DhM5lKe0_00002718 | 227_MU4DhM5lKe0_00002718.jpg | {
"data_source": "VIPSeg",
"file_name": "227_MU4DhM5lKe0_00002718.jpg",
"height": 720,
"id": "227_MU4DhM5lKe0_00002718",
"width": 1280
} | {
"caption": "In a home office setting, a man in white thermal upperwear carefully guides a large, freshly printed colour photograph of a waterfront as it emerges from a black Canon printer. Next to a cluttered back blue wall that features a white panelled door, shelving with a storage box, and some stereo equipment on a wooden table, while a small portion of the floor is visible below.",
"caption_ann": "In a home office setting, a <4:man in white thermal upperwear> carefully guides a <7:large, freshly printed colour photograph of a waterfront> as it emerges from a <8:black Canon printer>. Next to a <0:cluttered back blue wall> that features a <3:white panelled door>, shelving with a <5:storage box>, and some <2:stereo equipment> on a <6:wooden table>, while a small portion of the <1:floor> is visible below.",
"id": 1744,
"image_id": "227_MU4DhM5lKe0_00002718",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "man in white thermal upperwear"
},
{
"mask_ids": [
7
],
"txt_desc": "large, freshly printed colour photograph of a waterfront"
},
{
"mask_ids": [
8
],
"txt_desc": "black Canon printer"
},
{
"mask_ids": [
0
],
"txt_desc": "cluttered back blue wall"
},
{
"mask_ids": [
3
],
"txt_desc": "white panelled door"
},
{
"mask_ids": [
5
],
"txt_desc": "storage box"
},
{
"mask_ids": [
2
],
"txt_desc": "stereo equipment"
},
{
"mask_ids": [
6
],
"txt_desc": "wooden table"
},
{
"mask_ids": [
1
],
"txt_desc": "floor"
}
],
"labels": [
"wall",
"floor",
"other_electronic_product",
"door",
"person",
"box",
"table_or_desk",
"painting_or_poster",
"printer"
]
} | [
{
"area": 68609,
"bbox": [
126,
0,
1154,
720
],
"category_id": 0,
"id": 18835,
"image_id": "227_MU4DhM5lKe0_00002718",
"iscrowd": 0,
"segmentation": {
"counts": "llh22Xf06M3J7M2J6M3J6M3J6M3J7L3J6N2J:I4I9L202N002N0O3N002N002MO3N001PMb\\Ol2ac00001O1O001N10... | [
{
"id": 1,
"name": "object"
}
] |
train | 228_d8DVFvcsQtk_00001563 | 228_d8DVFvcsQtk_00001563.jpg | {
"data_source": "VIPSeg",
"file_name": "228_d8DVFvcsQtk_00001563.jpg",
"height": 720,
"id": "228_d8DVFvcsQtk_00001563",
"width": 1280
} | {
"caption": "A smiling man in a white graphic t-shirt and a smiling woman in a pink t-shirt are excitedly unboxing a new purchase together while sitting on a black sofa. The man lifts the black printer, still enclosed in its protective plastic bag, from the open cardboard box. The scene unfolds in a room with a cream-coloured wall and a dark curtain.",
"caption_ann": "A <3:smiling man in a white graphic t-shirt> and a <2:smiling woman in a pink t-shirt> are excitedly unboxing a new purchase together while sitting on a <6:black sofa>. The man lifts the <7:black printer>, still enclosed in its <5:protective plastic bag>, from the <4:open cardboard box>. The scene unfolds in a room with a <0:cream-coloured wall> and a <1:dark curtain>.",
"id": 1745,
"image_id": "228_d8DVFvcsQtk_00001563",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "smiling man in a white graphic t-shirt"
},
{
"mask_ids": [
2
],
"txt_desc": "smiling woman in a pink t-shirt"
},
{
"mask_ids": [
6
],
"txt_desc": "black sofa"
},
{
"mask_ids": [
7
],
"txt_desc": "black printer"
},
{
"mask_ids": [
5
],
"txt_desc": "protective plastic bag"
},
{
"mask_ids": [
4
],
"txt_desc": "open cardboard box"
},
{
"mask_ids": [
0
],
"txt_desc": "cream-coloured wall"
},
{
"mask_ids": [
1
],
"txt_desc": "dark curtain"
}
],
"labels": [
"wall",
"curtain",
"person",
"person",
"box",
"bag_or_package",
"sofa",
"printer"
]
} | [
{
"area": 180496,
"bbox": [
0,
0,
1280,
522
],
"category_id": 0,
"id": 18844,
"image_id": "228_d8DVFvcsQtk_00001563",
"iscrowd": 0,
"segmentation": {
"counts": "`0[2Uf0Z9fEU1K[Ke3eM[3[KejP5lNTYoJ1O3M1O3M1O3M1O3`^OZ1P;gNnDZ1R;iNhD[1W;fNfD\\1Z;gN`D\\1`;eN]... | [
{
"id": 1,
"name": "object"
}
] |
train | 228_dFj0gABzA0g_00003416 | 228_dFj0gABzA0g_00003416.jpg | {
"data_source": "VIPSeg",
"file_name": "228_dFj0gABzA0g_00003416.jpg",
"height": 720,
"id": "228_dFj0gABzA0g_00003416",
"width": 1280
} | {
"caption": "A yellow forklift drives on a concrete road next to the sea, as a person in red and a person in white tshirt walk along a concrete road, and a group of people,stand together, with a white shelf partially visible on the right.",
"caption_ann": "A <2:yellow forklift> drives on a <0:concrete road> next to the <1:sea>, as a <4:person in red> and a <5:person in white tshirt> walk along a <0:concrete road>, and a group of <6,7,8,9,10:people>,stand together, with a <3:white shelf> partially visible on the right.",
"id": 1746,
"image_id": "228_dFj0gABzA0g_00003416",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "yellow forklift"
},
{
"mask_ids": [
0
],
"txt_desc": "concrete road"
},
{
"mask_ids": [
1
],
"txt_desc": "sea"
},
{
"mask_ids": [
4
],
"txt_desc": "person in red"
},
{
"mask_ids": [
5
],
"txt_desc": "person in white tshirt"
},
{
"mask_ids": [
0
],
"txt_desc": "concrete road"
},
{
"mask_ids": [
6,
7,
8,
9,
10
],
"txt_desc": "people"
},
{
"mask_ids": [
3
],
"txt_desc": "white shelf"
}
],
"labels": [
"road",
"sea",
"wheeled_machine",
"shelf",
"person",
"person",
"person",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 550540,
"bbox": [
0,
0,
1280,
720
],
"category_id": 18,
"id": 18852,
"image_id": "228_dFj0gABzA0g_00003416",
"iscrowd": 0,
"segmentation": {
"counts": "ga0i4fa01O10000N20000O1O10000O100O100O1O100O1O10000O100O100O1O10000O100O100O1O10000O100O1000... | [
{
"id": 1,
"name": "object"
}
] |
train | 229_muaNgE7eiqo_00001158 | 229_muaNgE7eiqo_00001158.jpg | {
"data_source": "VIPSeg",
"file_name": "229_muaNgE7eiqo_00001158.jpg",
"height": 720,
"id": "229_muaNgE7eiqo_00001158",
"width": 1280
} | {
"caption": "A man with a black beard, wearing a light blue button-down shirt, gestures expressively as he speaks, positioned between a red beanbag chair on his left and a sofa with a modern circular pattern on his right. The light yellow wall behind him functions as a personal gallery, adorned with a large collection of eighteen small, framed black-and-white photos and three large framed, colourful art pieces. A vintage-style spotlight lamp on a wooden tripod stands nearby on the wood-panelled floor, completing the eclectic decor.",
"caption_ann": "A <4:man with a black beard, wearing a light blue button-down shirt,> gestures expressively as he speaks, positioned between a <5:red beanbag chair> on his left and a <6:sofa with a modern circular pattern> on his right. The <0:light yellow wall> behind him functions as a personal gallery, adorned with a large collection of <7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24:eighteen small, framed black-and-white photos> and <25,26,27:three large framed, colourful art pieces>. A <2:vintage-style spotlight lamp> on a <3:wooden tripod> stands nearby on the <1:wood-panelled floor>, completing the eclectic decor.",
"id": 1747,
"image_id": "229_muaNgE7eiqo_00001158",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "man with a black beard, wearing a light blue button-down shirt,"
},
{
"mask_ids": [
5
],
"txt_desc": "red beanbag chair"
},
{
"mask_ids": [
6
],
"txt_desc": "sofa with a modern circular pattern"
},
{
"mask_ids": [
0
],
"txt_desc": "light yellow wall"
},
{
"mask_ids": [
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24
],
"txt_desc": "eighteen small, framed black-and-white photos"
},
{
"mask_ids": [
25,
26,
27
],
"txt_desc": "three large framed, colourful art pieces"
},
{
"mask_ids": [
2
],
"txt_desc": "vintage-style spotlight lamp"
},
{
"mask_ids": [
3
],
"txt_desc": "wooden tripod"
},
{
"mask_ids": [
1
],
"txt_desc": "wood-panelled floor"
}
],
"labels": [
"wall",
"floor",
"lamp",
"shelf",
"person",
"sofa",
"sofa",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster",
"painting_or_poster"
]
} | [
{
"area": 519618,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18863,
"image_id": "229_muaNgE7eiqo_00001158",
"iscrowd": 0,
"segmentation": {
"counts": "0ba0n41O0000001O001O0000001O001O0000001O001O0000001O001O0000001O00001O0000001OO100O1O1O1O100O... | [
{
"id": 1,
"name": "object"
}
] |
train | 231_vd4jZBGxSQs_00000978 | 231_vd4jZBGxSQs_00000978.jpg | {
"data_source": "VIPSeg",
"file_name": "231_vd4jZBGxSQs_00000978.jpg",
"height": 720,
"id": "231_vd4jZBGxSQs_00000978",
"width": 1280
} | {
"caption": "A smiling man in a dark green T-shirt gestures enthusiastically in front of the camera. His high-tech setup is set against a light-coloured wall, which is dominated by a large computer monitor with space wallpaper and a smaller vertical screen that are placed on a white table. The white table is also adorned with a black adjustable lamp near a closed laptop, a mobile phone on a stand, a pixelated digital clock, and a small plant in a yellow pot besides a dark grey mat. On the right side of the image, a wooden storage cabinet that has a small potted plant completes the workspace, which sits on a tiled floor.",
"caption_ann": "A <6:smiling man in a dark green T-shirt> gestures enthusiastically in front of the camera. His high-tech setup is set against a <0:light-coloured wall>, which is dominated by a large <11:computer monitor with space wallpaper> and a <12:smaller vertical screen> that are placed on a <8:white table>. The <8:white table> is also adorned with a <3:black adjustable lamp> near a <13:closed laptop>, a <14:mobile phone on a stand>, a <7:pixelated digital clock>, and a <4:small plant> in a <10:yellow pot> besides a <2:dark grey mat>. On the right side of the image, a <5:wooden storage cabinet> that has a <9:small potted plant> completes the workspace, which sits on a <1:tiled floor>.",
"id": 1748,
"image_id": "231_vd4jZBGxSQs_00000978",
"label_matched": [
{
"mask_ids": [
6
],
"txt_desc": "smiling man in a dark green T-shirt"
},
{
"mask_ids": [
0
],
"txt_desc": "light-coloured wall"
},
{
"mask_ids": [
11
],
"txt_desc": "computer monitor with space wallpaper"
},
{
"mask_ids": [
12
],
"txt_desc": "smaller vertical screen"
},
{
"mask_ids": [
8
],
"txt_desc": "white table"
},
{
"mask_ids": [
8
],
"txt_desc": "white table"
},
{
"mask_ids": [
3
],
"txt_desc": "black adjustable lamp"
},
{
"mask_ids": [
13
],
"txt_desc": "closed laptop"
},
{
"mask_ids": [
14
],
"txt_desc": "mobile phone on a stand"
},
{
"mask_ids": [
7
],
"txt_desc": "pixelated digital clock"
},
{
"mask_ids": [
4
],
"txt_desc": "small plant"
},
{
"mask_ids": [
10
],
"txt_desc": "yellow pot"
},
{
"mask_ids": [
2
],
"txt_desc": "dark grey mat"
},
{
"mask_ids": [
5
],
"txt_desc": "wooden storage cabinet"
},
{
"mask_ids": [
9
],
"txt_desc": "small potted plant"
},
{
"mask_ids": [
1
],
"txt_desc": "tiled floor"
}
],
"labels": [
"wall",
"floor",
"cushion_or_carpet",
"lamp",
"other_plant",
"cupboard_or_showcase_or_storage_rack",
"person",
"box",
"table_or_desk",
"flower_pot_or_vase",
"flower_pot_or_vase",
"screen_or_television",
"screen_or_television",
"computer",
"Mobile_phone"
]
} | [
{
"area": 383242,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18891,
"image_id": "231_vd4jZBGxSQs_00000978",
"iscrowd": 0,
"segmentation": {
"counts": "0_Vj19]_VNP1VO4F;K5E;K5E:L5E;K5E;K5E;XIkKPKZ4o4iKdJb4[5`KaJc4^5`K`J`4^5bKbJ]4^5fK`JZ4_5hK`JX4... | [
{
"id": 1,
"name": "object"
}
] |
train | 232_ES2o87n0vqE_00001799 | 232_ES2o87n0vqE_00001799.jpg | {
"data_source": "VIPSeg",
"file_name": "232_ES2o87n0vqE_00001799.jpg",
"height": 720,
"id": "232_ES2o87n0vqE_00001799",
"width": 1280
} | {
"caption": "A child in a plaid shirt climbs a bright orange climbing wall filled with colorful handholds, with the ceiling visible at the top.",
"caption_ann": "A <2:child in a plaid shirt> climbs a <0:bright orange climbing wall> filled with colorful handholds, with the <1:ceiling> visible at the top.",
"id": 1749,
"image_id": "232_ES2o87n0vqE_00001799",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "child in a plaid shirt"
},
{
"mask_ids": [
0
],
"txt_desc": "bright orange climbing wall"
},
{
"mask_ids": [
1
],
"txt_desc": "ceiling"
}
],
"labels": [
"wall",
"ceiling",
"person"
]
} | [
{
"area": 843575,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18906,
"image_id": "232_ES2o87n0vqE_00001799",
"iscrowd": 0,
"segmentation": {
"counts": "e0ke0e000000000000000000000000000000000000000000000000000000001O000000O1000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 233_LkxwS2qrluA_00003258 | 233_LkxwS2qrluA_00003258.jpg | {
"data_source": "VIPSeg",
"file_name": "233_LkxwS2qrluA_00003258.jpg",
"height": 720,
"id": "233_LkxwS2qrluA_00003258",
"width": 1280
} | {
"caption": "In a softly lit office, a woman with curly hair, wearing glasses and a beaded necklace, speaks thoughtfully while seated next to a light-colored desk. The desk is neatly organized with a collection of books and binders, a small bouquet of flowers, and an illuminated table lamp with a white shade. The background of the room is composed of a pale, neutral-colored wall, a window covered by white blinds, and a sliver of the white ceiling visible above.",
"caption_ann": "In a softly lit office, a <6:woman with curly hair, wearing glasses and a beaded necklace,> speaks thoughtfully while seated next to a <7:light-colored desk>. The desk is neatly organized with a collection of <4:books and binders>, a <3:small bouquet of flowers>, and an illuminated <2:table lamp with a white shade>. The background of the room is composed of a <0:pale, neutral-colored wall>, a <5:window covered by white blinds>, and a sliver of the <1:white ceiling> visible above.",
"id": 1750,
"image_id": "233_LkxwS2qrluA_00003258",
"label_matched": [
{
"mask_ids": [
6
],
"txt_desc": "woman with curly hair, wearing glasses and a beaded necklace,"
},
{
"mask_ids": [
7
],
"txt_desc": "light-colored desk"
},
{
"mask_ids": [
4
],
"txt_desc": "books and binders"
},
{
"mask_ids": [
3
],
"txt_desc": "small bouquet of flowers"
},
{
"mask_ids": [
2
],
"txt_desc": "table lamp with a white shade"
},
{
"mask_ids": [
0
],
"txt_desc": "pale, neutral-colored wall"
},
{
"mask_ids": [
5
],
"txt_desc": "window covered by white blinds"
},
{
"mask_ids": [
1
],
"txt_desc": "white ceiling"
}
],
"labels": [
"wall",
"ceiling",
"lamp",
"flower",
"book",
"window",
"person",
"table_or_desk"
]
} | [
{
"area": 167434,
"bbox": [
0,
0,
1280,
432
],
"category_id": 0,
"id": 18909,
"image_id": "233_LkxwS2qrluA_00003258",
"iscrowd": 0,
"segmentation": {
"counts": "0`=P900O100O100O1O100O100O100O100O100O100O1O100O100O100O10000O1000000O10000O10000O1000000O100... | [
{
"id": 1,
"name": "object"
}
] |
train | 234_SMZ1WKnm8_I_00000166 | 234_SMZ1WKnm8_I_00000166.jpg | {
"data_source": "VIPSeg",
"file_name": "234_SMZ1WKnm8_I_00000166.jpg",
"height": 720,
"id": "234_SMZ1WKnm8_I_00000166",
"width": 1280
} | {
"caption": "A young girl wearing a white top and a pink tutu joyfully jumps mid-air with her arms outstretched on a yellow textured trampoline. In the background, a child in blue upperwear and another child with black bottomwear are also playing on the trampoline. The entire play area is enclosed by a red padded frame, with a larger indoor play fence visible.",
"caption_ann": "A <3:young girl wearing a white top and a pink tutu> joyfully jumps mid-air with her arms outstretched on a <2:yellow textured trampoline>. In the background, <4:a child in blue upperwear> and <5:another child with black bottomwear> are also playing on the trampoline. The entire play area is enclosed by a <1:red padded frame>, with a <0:larger indoor play fence> visible.",
"id": 1751,
"image_id": "234_SMZ1WKnm8_I_00000166",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "young girl wearing a white top and a pink tutu"
},
{
"mask_ids": [
2
],
"txt_desc": "yellow textured trampoline"
},
{
"mask_ids": [
4
],
"txt_desc": "a child in blue upperwear"
},
{
"mask_ids": [
5
],
"txt_desc": "another child with black bottomwear"
},
{
"mask_ids": [
1
],
"txt_desc": "red padded frame"
},
{
"mask_ids": [
0
],
"txt_desc": "larger indoor play fence"
}
],
"labels": [
"handrail_or_fence",
"pole",
"cushion_or_carpet",
"person",
"person",
"person"
]
} | [
{
"area": 284159,
"bbox": [
0,
0,
1280,
484
],
"category_id": 7,
"id": 18917,
"image_id": "234_SMZ1WKnm8_I_00000166",
"iscrowd": 0,
"segmentation": {
"counts": "0T2j6R4b0QL_Oj3b0[L^O_3c0fL]OU3c0PM\\Ok2e0[MZO_2g0fMXOU2i0PNWOk1i0ZNWO`1j0eNUOV1l0oNTOl0l0YOS... | [
{
"id": 1,
"name": "object"
}
] |
train | 237_fMEMjg_kQs4_00004079 | 237_fMEMjg_kQs4_00004079.jpg | {
"data_source": "VIPSeg",
"file_name": "237_fMEMjg_kQs4_00004079.jpg",
"height": 720,
"id": "237_fMEMjg_kQs4_00004079",
"width": 1280
} | {
"caption": "A man wearing a black shirt, camouflage shorts, sunglasses, and a hat stands in a dark lake with lily pads, holding a large fish with his hands, near green plants and grassy land in the background, with trees and a white lattice structure also visible.",
"caption_ann": "A <5:man wearing a black shirt, camouflage shorts, sunglasses, and a hat> stands in a <2:dark lake with lily pads>, holding a <6:large fish> with his hands, near <4:green plants> and <0:grassy land> in the background, with <3:trees> and a <1:white lattice structure> also visible.",
"id": 1752,
"image_id": "237_fMEMjg_kQs4_00004079",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "man wearing a black shirt, camouflage shorts, sunglasses, and a hat"
},
{
"mask_ids": [
2
],
"txt_desc": "dark lake with lily pads"
},
{
"mask_ids": [
6
],
"txt_desc": "large fish"
},
{
"mask_ids": [
4
],
"txt_desc": "green plants"
},
{
"mask_ids": [
0
],
"txt_desc": "grassy land"
},
{
"mask_ids": [
3
],
"txt_desc": "trees"
},
{
"mask_ids": [
1
],
"txt_desc": "white lattice structure"
}
],
"labels": [
"grass",
"other_construction",
"lake",
"tree",
"other_plant",
"person",
"other_animal"
]
} | [
{
"area": 202108,
"bbox": [
0,
40,
1280,
680
],
"category_id": 15,
"id": 18923,
"image_id": "237_fMEMjg_kQs4_00004079",
"iscrowd": 0,
"segmentation": {
"counts": "Z2g0je0O00000O1000O1O10000O1000O010000000000000000O10000O1000000000000O100O1000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 238_geUZ8UpkvZE_00002281 | 238_geUZ8UpkvZE_00002281.jpg | {
"data_source": "VIPSeg",
"file_name": "238_geUZ8UpkvZE_00002281.jpg",
"height": 720,
"id": "238_geUZ8UpkvZE_00002281",
"width": 1280
} | {
"caption": "A person in a red shirt, blue jeans, and a patterned hat stands on the gravelly ground beside a calm lake. He is fishing with a long fishing pole. In the background, a line of leafy trees on the far shore is visible under a pale sky.",
"caption_ann": "A <5:person in a red shirt, blue jeans, and a patterned hat> stands on the <1:gravelly ground> beside a <3:calm lake>. He is fishing with a <0:long fishing pole>. In the background, a <4:line of leafy trees> on the far shore is visible under a <2:pale sky>.",
"id": 1753,
"image_id": "238_geUZ8UpkvZE_00002281",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "person in a red shirt, blue jeans, and a patterned hat"
},
{
"mask_ids": [
1
],
"txt_desc": "gravelly ground"
},
{
"mask_ids": [
3
],
"txt_desc": "calm lake"
},
{
"mask_ids": [
0
],
"txt_desc": "long fishing pole"
},
{
"mask_ids": [
4
],
"txt_desc": "line of leafy trees"
},
{
"mask_ids": [
2
],
"txt_desc": "pale sky"
}
],
"labels": [
"pole",
"ground",
"sky",
"lake",
"tree",
"person"
]
} | [
{
"area": 644,
"bbox": [
417,
205,
113,
87
],
"category_id": 12,
"id": 18930,
"image_id": "238_geUZ8UpkvZE_00002281",
"iscrowd": 0,
"segmentation": {
"counts": "^_U91^f02O1N1O2N2O1N0001O01O0001O0001O01O000010O000001O0001O01O000010O000001O01O0001O00010O00... | [
{
"id": 1,
"name": "object"
}
] |
train | 245_9X2A2f6E5DI_00000917 | 245_9X2A2f6E5DI_00000917.jpg | {
"data_source": "VIPSeg",
"file_name": "245_9X2A2f6E5DI_00000917.jpg",
"height": 720,
"id": "245_9X2A2f6E5DI_00000917",
"width": 1280
} | {
"caption": "A black and yellow train emerges from a dense thicket of bare, wintery trees. Its bright headlights shining as it travels along an old, snow-dusted ground on a railroad track. A small patch of pale, overcast sky is visible at the top of the frame.",
"caption_ann": "A <4:black and yellow train> emerges from a <3:dense thicket of bare, wintery trees>. Its bright headlights shining as it travels along an <0:old, snow-dusted ground> on a <1:railroad track>. A small patch of <2:pale, overcast sky> is visible at the top of the frame.",
"id": 1754,
"image_id": "245_9X2A2f6E5DI_00000917",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "black and yellow train"
},
{
"mask_ids": [
3
],
"txt_desc": "dense thicket of bare, wintery trees"
},
{
"mask_ids": [
0
],
"txt_desc": "old, snow-dusted ground"
},
{
"mask_ids": [
1
],
"txt_desc": "railroad track"
},
{
"mask_ids": [
2
],
"txt_desc": "pale, overcast sky"
}
],
"labels": [
"rail",
"ground",
"sky",
"tree",
"train"
]
} | [
{
"area": 20573,
"bbox": [
0,
493,
989,
227
],
"category_id": 9,
"id": 18936,
"image_id": "245_9X2A2f6E5DI_00000917",
"iscrowd": 0,
"segmentation": {
"counts": "ad02^f09F6K2M10O100O0100O1000O10O01000O10O1000O100O1000000O100O0100000O100O1000000O100O010000... | [
{
"id": 1,
"name": "object"
}
] |
train | 24_mhz8I2-L7-E_00000505 | 24_mhz8I2-L7-E_00000505.jpg | {
"data_source": "VIPSeg",
"file_name": "24_mhz8I2-L7-E_00000505.jpg",
"height": 720,
"id": "24_mhz8I2-L7-E_00000505",
"width": 1280
} | {
"caption": "A young girl with long black hair, wearing a pink and white top over patterned leggings, stands looking forward with a serious expression. She is holding a small, black traveling case by its handle in her left hand. The room has a light-colored wood-plank floor and a beige wall in the background. Behind the girl, there is a closed white six-panel door with a brass doorknob. To her right, a large, light-colored curtain hangs down, and on the far right of the frame, a portion of a blue plastic storage drawer unit is visible.",
"caption_ann": "A <5:young girl with long black hair, wearing a pink and white top over patterned leggings>, stands looking forward with a serious expression. She is holding a <6:small, black traveling case> by its handle in her left hand. The room has a <1:light-colored wood-plank floor> and a <0:beige wall> in the background. Behind the <5:girl>, there is a closed <4:white six-panel door> with a brass doorknob. To her right, a large, <3:light-colored curtain> hangs down, and on the far right of the frame, a portion of a <2:blue plastic storage drawer unit> is visible.",
"id": 1755,
"image_id": "24_mhz8I2-L7-E_00000505",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "young girl with long black hair, wearing a pink and white top over patterned leggings"
},
{
"mask_ids": [
6
],
"txt_desc": "small, black traveling case"
},
{
"mask_ids": [
1
],
"txt_desc": "light-colored wood-plank floor"
},
{
"mask_ids": [
0
],
"txt_desc": "beige wall"
},
{
"mask_ids": [
5
],
"txt_desc": "girl"
},
{
"mask_ids": [
4
],
"txt_desc": "white six-panel door"
},
{
"mask_ids": [
3
],
"txt_desc": "light-colored curtain"
},
{
"mask_ids": [
2
],
"txt_desc": "blue plastic storage drawer unit"
}
],
"labels": [
"wall",
"floor",
"cupboard_or_showcase_or_storage_rack",
"curtain",
"door",
"person",
"traveling_case_or_trolley_case"
]
} | [
{
"area": 313887,
"bbox": [
480,
1,
800,
719
],
"category_id": 0,
"id": 18941,
"image_id": "24_mhz8I2-L7-E_00000505",
"iscrowd": 0,
"segmentation": {
"counts": "Q`a:6Zf0<D<D<D<D<D<D<D<D<D<D;E<D<D<D<D<D<D<D<D<D;E<D<D<D<D<D<D<D<D<D<D;E;E<D;E9G6J7I7I6J7I4LN... | [
{
"id": 1,
"name": "object"
}
] |
train | 250_gs5KlfuL1gI_00001248 | 250_gs5KlfuL1gI_00001248.jpg | {
"data_source": "VIPSeg",
"file_name": "250_gs5KlfuL1gI_00001248.jpg",
"height": 720,
"id": "250_gs5KlfuL1gI_00001248",
"width": 1280
} | {
"caption": "The image displays a man in a plaid jacket and a black fedora with a blue feather speaking expressively, addressing a group of guests. There are other people who can be seen in the background, including a woman in a pink dress, a woman in a white top, and two men in a dark suit, who are seated together on a light blue upholstered sofa at a long table decorated with a red floral arrangement and four wine glasses. The entire scene is set within an elegantly decorated room, featuring a cream-coloured wall and three windows adorned with elaborate purple and white curtains.",
"caption_ann": "The image displays a <6:man in a plaid jacket and a black fedora with a blue feather> speaking expressively, addressing a group of guests. There are other people who can be seen in the background, including a <10:woman in a pink dress>, a <9:woman in a white top>, and <7,8:two men in a dark suit>, who are seated together on a <16:light blue upholstered sofa> at a <15:long table> decorated with a <1:red floral arrangement> and <11,12,13,14:four wine glasses>. The entire scene is set within an elegantly decorated room, featuring a <0:cream-coloured wall> and <3,4,5:three windows> adorned with elaborate <2:purple and white curtains>.",
"id": 1756,
"image_id": "250_gs5KlfuL1gI_00001248",
"label_matched": [
{
"mask_ids": [
6
],
"txt_desc": "man in a plaid jacket and a black fedora with a blue feather"
},
{
"mask_ids": [
10
],
"txt_desc": "woman in a pink dress"
},
{
"mask_ids": [
9
],
"txt_desc": "woman in a white top"
},
{
"mask_ids": [
7,
8
],
"txt_desc": "two men in a dark suit"
},
{
"mask_ids": [
16
],
"txt_desc": "light blue upholstered sofa"
},
{
"mask_ids": [
15
],
"txt_desc": "long table"
},
{
"mask_ids": [
1
],
"txt_desc": "red floral arrangement"
},
{
"mask_ids": [
11,
12,
13,
14
],
"txt_desc": "four wine glasses"
},
{
"mask_ids": [
0
],
"txt_desc": "cream-coloured wall"
},
{
"mask_ids": [
3,
4,
5
],
"txt_desc": "three windows"
},
{
"mask_ids": [
2
],
"txt_desc": "purple and white curtains"
}
],
"labels": [
"wall",
"flower",
"curtain",
"window",
"window",
"window",
"person",
"person",
"person",
"person",
"person",
"bottle_or_cup",
"bottle_or_cup",
"bottle_or_cup",
"bottle_or_cup",
"table_or_desk",
"sofa"
]
} | [
{
"area": 67726,
"bbox": [
0,
1,
1280,
576
],
"category_id": 0,
"id": 18948,
"image_id": "250_gs5KlfuL1gI_00001248",
"iscrowd": 0,
"segmentation": {
"counts": "_=W1Ye00000000000O10000000000000000000O010000000000O100000000000000000O1000O10000000000O100000... | [
{
"id": 1,
"name": "object"
}
] |
train | 252_NqaAMx_4ZNw_00000017 | 252_NqaAMx_4ZNw_00000017.jpg | {
"data_source": "VIPSeg",
"file_name": "252_NqaAMx_4ZNw_00000017.jpg",
"height": 720,
"id": "252_NqaAMx_4ZNw_00000017",
"width": 1280
} | {
"caption": "A large, multi-storey waterfront house with many windows is situated on the edge of the calm, dark green water. The house is flanked by lush, green trees under a clear blue sky. A long wooden bridge supported by numerous thick wooden poles extends from the shore, while another small dock with a white handrail can be seen beside a patch of manicured green grass.",
"caption_ann": "A <3:large, multi-storey waterfront house with many windows> is situated on the edge of the <6:calm, dark green water>. The <3:house> is flanked by <7:lush, green trees> under a <5:clear blue sky>. A <4:long wooden bridge> supported by <1:numerous thick wooden poles> extends from the shore, while another small dock with a <0:white handrail> can be seen beside a patch of <2:manicured green grass>.",
"id": 1757,
"image_id": "252_NqaAMx_4ZNw_00000017",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "large, multi-storey waterfront house with many windows"
},
{
"mask_ids": [
6
],
"txt_desc": "calm, dark green water"
},
{
"mask_ids": [
3
],
"txt_desc": "house"
},
{
"mask_ids": [
7
],
"txt_desc": "lush, green trees"
},
{
"mask_ids": [
5
],
"txt_desc": "clear blue sky"
},
{
"mask_ids": [
4
],
"txt_desc": "long wooden bridge"
},
{
"mask_ids": [
1
],
"txt_desc": "numerous thick wooden poles"
},
{
"mask_ids": [
0
],
"txt_desc": "white handrail"
},
{
"mask_ids": [
2
],
"txt_desc": "manicured green grass"
}
],
"labels": [
"handrail_or_fence",
"pole",
"grass",
"house",
"bridge",
"sky",
"water",
"tree"
]
} | [
{
"area": 8580,
"bbox": [
0,
421,
141,
110
],
"category_id": 7,
"id": 18965,
"image_id": "252_NqaAMx_4ZNw_00000017",
"iscrowd": 0,
"segmentation": {
"counts": "]>V2Zd00000O100000O010000000000O0100000O10000000000O1000000O10O10000000O10O1000O10000000000O10... | [
{
"id": 1,
"name": "object"
}
] |
train | 253_kgCzdbJVaGM_00001226 | 253_kgCzdbJVaGM_00001226.jpg | {
"data_source": "VIPSeg",
"file_name": "253_kgCzdbJVaGM_00001226.jpg",
"height": 720,
"id": "253_kgCzdbJVaGM_00001226",
"width": 1280
} | {
"caption": "A person drives a wooden speedboat on a calm river, with grassy land and many trees lining the shore. A house is visible on the right,along with a boathouse and a dock where four chairs are set up.",
"caption_ann": "A <5:person> drives a <4:wooden speedboat> on a <2:calm river>, with <0:grassy land> and <3:many trees> lining the shore. A <1:house> is visible on the right,along with a <1:boathouse and a dock> where <6,7,8,9:four chairs> are set up.",
"id": 1758,
"image_id": "253_kgCzdbJVaGM_00001226",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "person"
},
{
"mask_ids": [
4
],
"txt_desc": "wooden speedboat"
},
{
"mask_ids": [
2
],
"txt_desc": "calm river"
},
{
"mask_ids": [
0
],
"txt_desc": "grassy land"
},
{
"mask_ids": [
3
],
"txt_desc": "many trees"
},
{
"mask_ids": [
1
],
"txt_desc": "house"
},
{
"mask_ids": [
1
],
"txt_desc": "boathouse and a dock"
},
{
"mask_ids": [
6,
7,
8,
9
],
"txt_desc": "four chairs"
}
],
"labels": [
"grass",
"house",
"river",
"tree",
"ship_or_boat",
"person",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat"
]
} | [
{
"area": 61403,
"bbox": [
63,
195,
1184,
147
],
"category_id": 15,
"id": 18973,
"image_id": "253_kgCzdbJVaGM_00001226",
"iscrowd": 0,
"segmentation": {
"counts": "Wa\\11_f00O1000000O100000000O2O000O1000001O1O001O1OO10O10000001O00000001O001O1O0O1O0100O10... | [
{
"id": 1,
"name": "object"
}
] |
train | 255_EjctWA0PIDg_00000243 | 255_EjctWA0PIDg_00000243.jpg | {
"data_source": "VIPSeg",
"file_name": "255_EjctWA0PIDg_00000243.jpg",
"height": 720,
"id": "255_EjctWA0PIDg_00000243",
"width": 1280
} | {
"caption": "This image features an opening of a wooden structure, where a group of four people are walking towards the viewer on the ground. The group is comprised of a woman with blonde hair wearing a sky blue cardigan, a man in a black t-shirt, a man in a white apron over his clothes, and a man in a cap and brown vest. In the background, a rustic building with a tiled roof on a grass patch that is backed by lush green trees can be seen.",
"caption_ann": "This image features an opening of a <0:wooden structure>, where a group of four people are walking towards the viewer on the <1:ground>. The group is comprised of a <5:woman with blonde hair wearing a sky blue cardigan>, a <6:man in a black t-shirt>, a <7:man in a white apron over his clothes>, and a <8:man in a cap and brown vest>. In the background, a <3:rustic building with a tiled roof> on a <2:grass patch> that is backed by <4:lush green trees> can be seen.",
"id": 1759,
"image_id": "255_EjctWA0PIDg_00000243",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "wooden structure"
},
{
"mask_ids": [
1
],
"txt_desc": "ground"
},
{
"mask_ids": [
5
],
"txt_desc": "woman with blonde hair wearing a sky blue cardigan"
},
{
"mask_ids": [
6
],
"txt_desc": "man in a black t-shirt"
},
{
"mask_ids": [
7
],
"txt_desc": "man in a white apron over his clothes"
},
{
"mask_ids": [
8
],
"txt_desc": "man in a cap and brown vest"
},
{
"mask_ids": [
3
],
"txt_desc": "rustic building with a tiled roof"
},
{
"mask_ids": [
2
],
"txt_desc": "grass patch"
},
{
"mask_ids": [
4
],
"txt_desc": "lush green trees"
}
],
"labels": [
"wall",
"ground",
"grass",
"other_construction",
"tree",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 475150,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 18983,
"image_id": "255_EjctWA0PIDg_00000243",
"iscrowd": 0,
"segmentation": {
"counts": "Y;W;1X;00000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 256_F8__hNZTKCw_00003192 | 256_F8__hNZTKCw_00003192.jpg | {
"data_source": "VIPSeg",
"file_name": "256_F8__hNZTKCw_00003192.jpg",
"height": 720,
"id": "256_F8__hNZTKCw_00003192",
"width": 1280
} | {
"caption": "A man wearing sunglasses and a gray short-sleeved shirt sits with his legs spread out in the open back of a black van, with dark ground, grass, and trees visible around the van, all under a clear sky.",
"caption_ann": "A <5:man wearing sunglasses and a gray short-sleeved shirt> sits with his legs spread out in the open back of a <4:black van>, with <0:dark ground>, <1:grass>, and <3:trees> visible around the <4:van>, all under a <2:clear sky>.",
"id": 1760,
"image_id": "256_F8__hNZTKCw_00003192",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "man wearing sunglasses and a gray short-sleeved shirt"
},
{
"mask_ids": [
4
],
"txt_desc": "black van"
},
{
"mask_ids": [
0
],
"txt_desc": "dark ground"
},
{
"mask_ids": [
1
],
"txt_desc": "grass"
},
{
"mask_ids": [
3
],
"txt_desc": "trees"
},
{
"mask_ids": [
4
],
"txt_desc": "van"
},
{
"mask_ids": [
2
],
"txt_desc": "clear sky"
}
],
"labels": [
"ground",
"grass",
"sky",
"tree",
"car",
"person"
]
} | [
{
"area": 35331,
"bbox": [
0,
593,
1029,
127
],
"category_id": 14,
"id": 18992,
"image_id": "256_F8__hNZTKCw_00003192",
"iscrowd": 0,
"segmentation": {
"counts": "ab0o3bb0O00000000000000O1001O000000000000000000000000001O0000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 257_wBypuJSdrPk_00001742 | 257_wBypuJSdrPk_00001742.jpg | {
"data_source": "VIPSeg",
"file_name": "257_wBypuJSdrPk_00001742.jpg",
"height": 720,
"id": "257_wBypuJSdrPk_00001742",
"width": 1280
} | {
"caption": "A man in a blue jacket and jeans working on the exterior of a brick house while standing carefully on a yellow stepladder which is set up on the paved ground. In the bottom right corner of the image, a leafy green bush and a black car are visible, while in the top right corner, a silver sky is visible.",
"caption_ann": "A <6:man in a blue jacket and jeans> working on the exterior of a <1:brick house> while standing carefully on a <4:yellow stepladder> which is set up on the <0:paved ground>. In the bottom right corner of the image, a <3:leafy green bush> and a <5:black car> are visible, while in the top right corner, a <2:silver sky> is visible.",
"id": 1761,
"image_id": "257_wBypuJSdrPk_00001742",
"label_matched": [
{
"mask_ids": [
6
],
"txt_desc": "man in a blue jacket and jeans"
},
{
"mask_ids": [
1
],
"txt_desc": "brick house"
},
{
"mask_ids": [
4
],
"txt_desc": "yellow stepladder"
},
{
"mask_ids": [
0
],
"txt_desc": "paved ground"
},
{
"mask_ids": [
3
],
"txt_desc": "leafy green bush"
},
{
"mask_ids": [
5
],
"txt_desc": "black car"
},
{
"mask_ids": [
2
],
"txt_desc": "silver sky"
}
],
"labels": [
"ground",
"house",
"sky",
"other_plant",
"ladder",
"car",
"person"
]
} | [
{
"area": 19919,
"bbox": [
681,
349,
599,
371
],
"category_id": 14,
"id": 18998,
"image_id": "257_wBypuJSdrPk_00001742",
"iscrowd": 0,
"segmentation": {
"counts": "o`o>1_f0000O1O100O100O100O100O100O100O100O100O100O1O10000O1O10000O1O100O1O10000O1O10000O1O... | [
{
"id": 1,
"name": "object"
}
] |
train | 267_Xy5ESqoMEbs_00000047 | 267_Xy5ESqoMEbs_00000047.jpg | {
"data_source": "VIPSeg",
"file_name": "267_Xy5ESqoMEbs_00000047.jpg",
"height": 720,
"id": "267_Xy5ESqoMEbs_00000047",
"width": 1280
} | {
"caption": "A young girl, wearing a colourful swimsuit and a life jacket, stands at the edge of a wooden dock, looking down with a hesitant expression at the calm lake. A blue inflatable water toy and a covered black motorboat are floating on water, while another partially visible boat rests on the dock in the bottom right corner. The scene is set against a backdrop of mountains covered with lush green trees, all under a partly cloudy sky.",
"caption_ann": "A <8:young girl, wearing a colourful swimsuit and a life jacket,> stands at the edge of a <0:wooden dock>, looking down with a hesitant expression at the <3:calm lake>. A <5:blue inflatable water toy> and a <6:covered black motorboat> are floating on water, while another <7:partially visible boat> rests on the <0:dock> in the bottom right corner. The scene is set against a backdrop of <2:mountains> covered with <4:lush green trees>, all under a <1:partly cloudy sky>.",
"id": 1762,
"image_id": "267_Xy5ESqoMEbs_00000047",
"label_matched": [
{
"mask_ids": [
8
],
"txt_desc": "young girl, wearing a colourful swimsuit and a life jacket,"
},
{
"mask_ids": [
0
],
"txt_desc": "wooden dock"
},
{
"mask_ids": [
3
],
"txt_desc": "calm lake"
},
{
"mask_ids": [
5
],
"txt_desc": "blue inflatable water toy"
},
{
"mask_ids": [
6
],
"txt_desc": "covered black motorboat"
},
{
"mask_ids": [
7
],
"txt_desc": "partially visible boat"
},
{
"mask_ids": [
0
],
"txt_desc": "dock"
},
{
"mask_ids": [
2
],
"txt_desc": "mountains"
},
{
"mask_ids": [
4
],
"txt_desc": "lush green trees"
},
{
"mask_ids": [
1
],
"txt_desc": "partly cloudy sky"
}
],
"labels": [
"bridge",
"sky",
"mountain",
"lake",
"tree",
"ship_or_boat",
"ship_or_boat",
"ship_or_boat",
"person"
]
} | [
{
"area": 77931,
"bbox": [
0,
527,
728,
193
],
"category_id": 23,
"id": 19005,
"image_id": "267_Xy5ESqoMEbs_00000047",
"iscrowd": 0,
"segmentation": {
"counts": "Yf07\\e0m0A?H80000O100O100000000O100O1000000O100O1000000O100O1000000O100O100000000O100O10000... | [
{
"id": 1,
"name": "object"
}
] |
train | 268_cdTD0htRdf8_00000017 | 268_cdTD0htRdf8_00000017.jpg | {
"data_source": "VIPSeg",
"file_name": "268_cdTD0htRdf8_00000017.jpg",
"height": 720,
"id": "268_cdTD0htRdf8_00000017",
"width": 1280
} | {
"caption": "The image displays three children standing on a wooden dock next to a boat with a blue top. The child on the left is holding a yellow and red surfboard in his right hand. The child wearing blue life vest is standing in the middle and holding a small white surfboard. A partially visible child on the right is also holding a black surfboard. The dock is surrounded by a vast lake with a tree-covered shore in the background under a blue sky.",
"caption_ann": "The image displays <5,6,7:three children> standing on a <0:wooden dock> next to a <4:boat with a blue top>. The <7:child> on the left is holding a <8:yellow and red surfboard> in his right hand. The <6:child wearing blue life vest> is standing in the middle and holding a <9:small white surfboard>. A <5:partially visible child> on the right is also holding a <10:black surfboard>. The <0:dock> is surrounded by a vast <2:lake> with a <3:tree-covered shore> in the background under a <1:blue sky>.",
"id": 1763,
"image_id": "268_cdTD0htRdf8_00000017",
"label_matched": [
{
"mask_ids": [
5,
6,
7
],
"txt_desc": "three children"
},
{
"mask_ids": [
0
],
"txt_desc": "wooden dock"
},
{
"mask_ids": [
4
],
"txt_desc": "boat with a blue top"
},
{
"mask_ids": [
7
],
"txt_desc": "child"
},
{
"mask_ids": [
8
],
"txt_desc": "yellow and red surfboard"
},
{
"mask_ids": [
6
],
"txt_desc": "child wearing blue life vest"
},
{
"mask_ids": [
9
],
"txt_desc": "small white surfboard"
},
{
"mask_ids": [
5
],
"txt_desc": "partially visible child"
},
{
"mask_ids": [
10
],
"txt_desc": "black surfboard"
},
{
"mask_ids": [
0
],
"txt_desc": "dock"
},
{
"mask_ids": [
2
],
"txt_desc": "lake"
},
{
"mask_ids": [
3
],
"txt_desc": "tree-covered shore"
},
{
"mask_ids": [
1
],
"txt_desc": "blue sky"
}
],
"labels": [
"bridge",
"sky",
"lake",
"tree",
"ship_or_boat",
"person",
"person",
"person",
"skateboard",
"skateboard",
"skateboard"
]
} | [
{
"area": 127364,
"bbox": [
538,
308,
742,
322
],
"category_id": 23,
"id": 19014,
"image_id": "268_cdTD0htRdf8_00000017",
"iscrowd": 0,
"segmentation": {
"counts": "Vej;4\\f0004L0N6L004L0N6L0000002N0N20002N0N4N00000N8J00^1bN00X1hN000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 271_qlb6M-vAmeo_00001287 | 271_qlb6M-vAmeo_00001287.jpg | {
"data_source": "VIPSeg",
"file_name": "271_qlb6M-vAmeo_00001287.jpg",
"height": 720,
"id": "271_qlb6M-vAmeo_00001287",
"width": 1280
} | {
"caption": "This image displays a several houses with gray roofs situated alongside a cracked road. A white car is driving around a corner, three cars are parked in front and numerous other cars are also visible in the background. Green lawns and a dry dirt grounds surround the houses with a concrete path running along them. Utility poles stand by the road and trees visible around the area.",
"caption_ann": "This image displays a <5:several houses with gray roofs> situated alongside a <3:cracked road>. A <10: white car> is driving around a corner, <7,8,9:three cars> are parked in front and <11,12,13,14,15,16,17,18:numerous other cars> are also visible in the background. <2:Green lawns> and a <1:dry dirt grounds> surround the <5:houses> with a <4:concrete path> running along them. <0:Utility poles> stand by the <3:road> and <6:trees> visible around the area.",
"id": 1764,
"image_id": "271_qlb6M-vAmeo_00001287",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "several houses with gray roofs"
},
{
"mask_ids": [
3
],
"txt_desc": "cracked road"
},
{
"mask_ids": [
10
],
"txt_desc": "white car"
},
{
"mask_ids": [
7,
8,
9
],
"txt_desc": "three cars"
},
{
"mask_ids": [
11,
12,
13,
14,
15,
16,
17,
18
],
"txt_desc": "numerous other cars"
},
{
"mask_ids": [
2
],
"txt_desc": "Green lawns"
},
{
"mask_ids": [
1
],
"txt_desc": "dry dirt grounds"
},
{
"mask_ids": [
5
],
"txt_desc": "houses"
},
{
"mask_ids": [
4
],
"txt_desc": "concrete path"
},
{
"mask_ids": [
0
],
"txt_desc": "Utility poles"
},
{
"mask_ids": [
3
],
"txt_desc": "road"
},
{
"mask_ids": [
6
],
"txt_desc": "trees"
}
],
"labels": [
"pole",
"ground",
"grass",
"road",
"path",
"house",
"tree",
"car",
"car",
"car",
"car",
"car",
"car",
"car",
"car",
"car",
"car",
"car",
"car"
]
} | [
{
"area": 5905,
"bbox": [
909,
147,
204,
573
],
"category_id": 12,
"id": 19025,
"image_id": "271_qlb6M-vAmeo_00001287",
"iscrowd": 0,
"segmentation": {
"counts": "gboc07Uf04I7N3L3N2M3N1N2O2N1M2O2NK6O11N0O11N2000O1N10M3010N2N100001001O0OL4O1101O0O10O10O1N... | [
{
"id": 1,
"name": "object"
}
] |
train | 275_1mhGD_1o71M_00007810 | 275_1mhGD_1o71M_00007810.jpg | {
"data_source": "VIPSeg",
"file_name": "275_1mhGD_1o71M_00007810.jpg",
"height": 720,
"id": "275_1mhGD_1o71M_00007810",
"width": 1280
} | {
"caption": "A man wearing a black shirt, black pants, and a grey backpack is riding an electric bike on a concrete road. A a concrete railing runs along the side of the concrete road on the right. A row of dense trees is visible beside the road, and a black pole stands in the top right corner of the image.",
"caption_ann": "A <5:man wearing a black shirt, black pants, and a grey backpack> is riding an <4:electric bike> on a <2:concrete road>. A <0:a concrete railing> runs along the side of the <2:concrete road> on the right. A row of <3:dense trees> is visible beside the <2:road>, and a <1:black pole> stands in the top right corner of the image.",
"id": 1765,
"image_id": "275_1mhGD_1o71M_00007810",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "man wearing a black shirt, black pants, and a grey backpack"
},
{
"mask_ids": [
4
],
"txt_desc": "electric bike"
},
{
"mask_ids": [
2
],
"txt_desc": "concrete road"
},
{
"mask_ids": [
0
],
"txt_desc": "a concrete railing"
},
{
"mask_ids": [
2
],
"txt_desc": "concrete road"
},
{
"mask_ids": [
3
],
"txt_desc": "dense trees"
},
{
"mask_ids": [
2
],
"txt_desc": "road"
},
{
"mask_ids": [
1
],
"txt_desc": "black pole"
}
],
"labels": [
"handrail_or_fence",
"pole",
"road",
"tree",
"bicycle",
"person"
]
} | [
{
"area": 124773,
"bbox": [
17,
12,
1242,
387
],
"category_id": 7,
"id": 19044,
"image_id": "275_1mhGD_1o71M_00007810",
"iscrowd": 0,
"segmentation": {
"counts": "ln;>Rf0n0SOl0TOh0WO10O10O0100O10O0100O010O010O01O001O0010O01O001O010O001O001O010O001O0010O0... | [
{
"id": 1,
"name": "object"
}
] |
train | 277_4dotp6yiju8_00003040 | 277_4dotp6yiju8_00003040.jpg | {
"data_source": "VIPSeg",
"file_name": "277_4dotp6yiju8_00003040.jpg",
"height": 720,
"id": "277_4dotp6yiju8_00003040",
"width": 1280
} | {
"caption": "In a classroom with a light green wall in the background, a woman with dark hair wearing a burnt-orange top seated on a wooden chair looks into a large white plastic bag with red lettering that is on a white wooden desk in front of her. Seated next to her on a wooden chair is a woman with her hair in a ponytail, wearing a white \"1990s\" t-shirt, who points a finger and speaks to her. The room is filled with other students, including a young man in a black t-shirt sitting at a and a desk on a chair which is next to a brown wooden door while a young man wearing gray shirt is reading a red book next to a girl in a black framed specs. In the background, a young man with a faded haircut is seated while talking to another young guy, and in the far right corner of the image, a girl wearing a watch looks down at a silver mobile phone, while another partially visible person can be seen next to her. The students are all sitting at five other wooden desks on five wooden chairs.",
"caption_ann": "In a classroom with a <0:light green wall> in the background, <4:a woman with dark hair wearing a burnt-orange top> seated on a <22:wooden chair> looks into a <11:large white plastic bag with red lettering> that is on a <16:white wooden desk> in front of her. Seated next to her on a <21:wooden chair> is a <5:woman with her hair in a ponytail, wearing a white \"1990s\" t-shirt>, who points a finger and speaks to her. The room is filled with other students, including a <3:young man in a black t-shirt> sitting at a and a <17:desk> on a <23:chair> which is next to a <2:brown wooden door> while a <9:young man wearing gray shirt> is reading a <1:red book> next to a <7:girl in a black framed specs>. In the background, a <6:young man with a faded haircut> is seated while talking to another <3:young guy>, and in the far right corner of the image, a <8:girl wearing a watch> looks down at a <25:silver mobile phone>, while another <10:partially visible person> can be seen next to her. The students are all sitting at <12,13,14,15,18:five other wooden desks> on <19,20,21,23,24:five wooden chairs>.",
"id": 1766,
"image_id": "277_4dotp6yiju8_00003040",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "light green wall"
},
{
"mask_ids": [
4
],
"txt_desc": "a woman with dark hair wearing a burnt-orange top"
},
{
"mask_ids": [
22
],
"txt_desc": "wooden chair"
},
{
"mask_ids": [
11
],
"txt_desc": "large white plastic bag with red lettering"
},
{
"mask_ids": [
16
],
"txt_desc": "white wooden desk"
},
{
"mask_ids": [
21
],
"txt_desc": "wooden chair"
},
{
"mask_ids": [
5
],
"txt_desc": "woman with her hair in a ponytail, wearing a white \"1990s\" t-shirt"
},
{
"mask_ids": [
3
],
"txt_desc": "young man in a black t-shirt"
},
{
"mask_ids": [
17
],
"txt_desc": "desk"
},
{
"mask_ids": [
23
],
"txt_desc": "chair"
},
{
"mask_ids": [
2
],
"txt_desc": "brown wooden door"
},
{
"mask_ids": [
9
],
"txt_desc": "young man wearing gray shirt"
},
{
"mask_ids": [
1
],
"txt_desc": "red book"
},
{
"mask_ids": [
7
],
"txt_desc": "girl in a black framed specs"
},
{
"mask_ids": [
6
],
"txt_desc": "young man with a faded haircut"
},
{
"mask_ids": [
3
],
"txt_desc": "young guy"
},
{
"mask_ids": [
8
],
"txt_desc": "girl wearing a watch"
},
{
"mask_ids": [
25
],
"txt_desc": "silver mobile phone"
},
{
"mask_ids": [
10
],
"txt_desc": "partially visible person"
},
{
"mask_ids": [
12,
13,
14,
15,
18
],
"txt_desc": "five other wooden desks"
},
{
"mask_ids": [
19,
20,
21,
23,
24
],
"txt_desc": "five wooden chairs"
}
],
"labels": [
"wall",
"book",
"door",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"bag_or_package",
"table_or_desk",
"table_or_desk",
"table_or_desk",
"table_or_desk",
"table_or_desk",
"table_or_desk",
"table_or_desk",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"Mobile_phone"
]
} | [
{
"area": 322531,
"bbox": [
0,
0,
1280,
521
],
"category_id": 0,
"id": 19050,
"image_id": "277_4dotp6yiju8_00003040",
"iscrowd": 0,
"segmentation": {
"counts": "0a<o91O001O00000000001O0000001O000000VO[FiDe9n:jFlDV9f:bGQE]8n:c1M3O1M3O1M3O1M3O100O1O100N200... | [
{
"id": 1,
"name": "object"
}
] |
train | 277_d4_yQugT37w_00000032 | 277_d4_yQugT37w_00000032.jpg | {
"data_source": "VIPSeg",
"file_name": "277_d4_yQugT37w_00000032.jpg",
"height": 720,
"id": "277_d4_yQugT37w_00000032",
"width": 1280
} | {
"caption": "The image features a vast, patchy green and brown grass driving range where a basket of white golf balls is placed. The central focus is on a group of three men. One man dressed in a black striped long-sleeved shirt and a matching black cap is selecting a golf club from his blue and white golf bag. Standing next to him and looking on is a man in a grey sweatshirt and a white cap. Another man wearing a grey jacket is carrying a bright turquoise backpack stands slightly behind them, observing the interaction. In the foreground, the backs of several spectators are visible, including a person in a dark blue jacket, part of a person wearing a white striped top, and the shoulder of another person in the bottom left corner.",
"caption_ann": "The image features a <0:vast, patchy green and brown grass driving range> where a basket of <8: white golf balls> is placed. The central focus is on a group of <2,3,4:three men>. One <2:man dressed in a black striped long-sleeved shirt and a matching black cap> is selecting a <1:golf club> from his <9:blue and white golf bag>. Standing next to him and looking on is a <3:man in a grey sweatshirt and a white cap>. Another <4:man wearing a grey jacket> is carrying a <10:bright turquoise backpack> stands slightly behind them, observing the interaction. In the foreground, the backs of several spectators are visible, including a <6:person in a dark blue jacket>, part of a <5:person wearing a white striped top>, and the shoulder of another <7:person> in the bottom left corner.",
"id": 1767,
"image_id": "277_d4_yQugT37w_00000032",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "vast, patchy green and brown grass driving range"
},
{
"mask_ids": [
8
],
"txt_desc": "white golf balls"
},
{
"mask_ids": [
2,
3,
4
],
"txt_desc": "three men"
},
{
"mask_ids": [
2
],
"txt_desc": "man dressed in a black striped long-sleeved shirt and a matching black cap"
},
{
"mask_ids": [
1
],
"txt_desc": "golf club"
},
{
"mask_ids": [
9
],
"txt_desc": "blue and white golf bag"
},
{
"mask_ids": [
3
],
"txt_desc": "man in a grey sweatshirt and a white cap"
},
{
"mask_ids": [
4
],
"txt_desc": "man wearing a grey jacket"
},
{
"mask_ids": [
10
],
"txt_desc": "bright turquoise backpack"
},
{
"mask_ids": [
6
],
"txt_desc": "person in a dark blue jacket"
},
{
"mask_ids": [
5
],
"txt_desc": "person wearing a white striped top"
},
{
"mask_ids": [
7
],
"txt_desc": "person"
}
],
"labels": [
"grass",
"bat",
"person",
"person",
"person",
"person",
"person",
"person",
"ball",
"bag_or_package",
"bag_or_package"
]
} | [
{
"area": 763071,
"bbox": [
0,
0,
1280,
720
],
"category_id": 15,
"id": 19076,
"image_id": "277_d4_yQugT37w_00000032",
"iscrowd": 0,
"segmentation": {
"counts": "0lc0d21O1O1O2N1O1O1O1O2N1O1O1O1O2N1O1O6J;E;E:F;E;EgSY21SlfM7I7L4K4J5N3L3K5K5N2M3N2M3N2N2O1M3... | [
{
"id": 1,
"name": "object"
}
] |
train | 278_5nXBeOiHXXo_00000505 | 278_5nXBeOiHXXo_00000505.jpg | {
"data_source": "VIPSeg",
"file_name": "278_5nXBeOiHXXo_00000505.jpg",
"height": 720,
"id": "278_5nXBeOiHXXo_00000505",
"width": 1280
} | {
"caption": "In a suburban backyard, two men are tending to a temporary pool. A man wearing a long-sleeved shirt, a cap, and sunglasses is bent over an open, white cooler, looking at its contents. Standing beside him, a man in a t-shirt and a backwards cap observes. They are standing on a patch of green grass next to a large, above-ground pool with murky green water. The backyard is enclosed by a peach-colored stucco wall and a pool safety fence. In the background, across a lake, the back of a house is visible. The scene is set under a bright, partly cloudy sky, with palm trees adding to the tropical feel. A pool hose is draped over the temporary pool, and a pair of oar can be seen standing along the wall. The paved deck of pool and steps are also visible. The large silver metallic pole is partially visible.",
"caption_ann": "In a suburban backyard, <14,15: two men> are tending to a temporary pool. A <14:man wearing a long-sleeved shirt, a cap, and sunglasses> is bent over an open, <16:white cooler>, looking at its contents. Standing beside him, a <15:man in a t-shirt and a backwards cap> observes. They are standing on a <4:patch of green grass> next to a <13:large, above-ground pool> with <9:murky green water>. The backyard is enclosed by a <0:peach-colored stucco wall> and a <1:pool safety fence>. In the background, across a <5:lake>, the back of a <7:house> is visible. The scene is set under a <8:bright, partly cloudy sky>, with <11:palm trees> adding to the tropical feel. A <10:pool hose> is draped over the <13:temporary pool>, and a <2: pair of oar> can be seen standing along the wall. The <3:paved deck of pool> and <6:steps> are also visible. The <12:large silver metallic pole> is partially visible.",
"id": 1768,
"image_id": "278_5nXBeOiHXXo_00000505",
"label_matched": [
{
"mask_ids": [
14,
15
],
"txt_desc": "two men"
},
{
"mask_ids": [
14
],
"txt_desc": "man wearing a long-sleeved shirt, a cap, and sunglasses"
},
{
"mask_ids": [
16
],
"txt_desc": "white cooler"
},
{
"mask_ids": [
15
],
"txt_desc": "man in a t-shirt and a backwards cap"
},
{
"mask_ids": [
4
],
"txt_desc": "patch of green grass"
},
{
"mask_ids": [
13
],
"txt_desc": "large, above-ground pool"
},
{
"mask_ids": [
9
],
"txt_desc": "murky green water"
},
{
"mask_ids": [
0
],
"txt_desc": "peach-colored stucco wall"
},
{
"mask_ids": [
1
],
"txt_desc": "pool safety fence"
},
{
"mask_ids": [
5
],
"txt_desc": "lake"
},
{
"mask_ids": [
7
],
"txt_desc": "house"
},
{
"mask_ids": [
8
],
"txt_desc": "bright, partly cloudy sky"
},
{
"mask_ids": [
11
],
"txt_desc": "palm trees"
},
{
"mask_ids": [
10
],
"txt_desc": "pool hose"
},
{
"mask_ids": [
13
],
"txt_desc": "temporary pool"
},
{
"mask_ids": [
2
],
"txt_desc": "pair of oar"
},
{
"mask_ids": [
3
],
"txt_desc": "paved deck of pool"
},
{
"mask_ids": [
6
],
"txt_desc": "steps"
},
{
"mask_ids": [
12
],
"txt_desc": "large silver metallic pole"
}
],
"labels": [
"wall",
"handrail_or_fence",
"pole",
"ground",
"grass",
"road",
"path",
"house",
"sky",
"water",
"pipeline",
"tree",
"shelf",
"bathtub",
"person",
"person",
"traveling_case_or_trolley_case"
]
} | [
{
"area": 259929,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 19087,
"image_id": "278_5nXBeOiHXXo_00000505",
"iscrowd": 0,
"segmentation": {
"counts": "0Q7`0aKm;`4RD`Kn;`4RD_Ko;a4QD_Ko;a4QD^KP<b4PD^KP<b4PD^KP<b4PD]KQ<c4oC\\KR<d4nC\\KR<d4nC\\KR<d... | [
{
"id": 1,
"name": "object"
}
] |
train | 27_97-FeRYjxXU_00001472 | 27_97-FeRYjxXU_00001472.jpg | {
"data_source": "VIPSeg",
"file_name": "27_97-FeRYjxXU_00001472.jpg",
"height": 720,
"id": "27_97-FeRYjxXU_00001472",
"width": 1280
} | {
"caption": "The image captures a lively scene at an indoor pool with a patterned tile floor. A group of five children are playing in the illuminated small pool. A girl with pigtails, wearing a black swimsuit with pink straps, has her back to the camera and is holding onto the pool's edge, near a curved metal handrail. A smiling boy is in the center of the pool, and three other children are also visible in the water. In the background, a pink and white patterned backpack rests on the floor, and there is a seating area where several five people are gathered around a table on five chairs. A purple and blue inflatable toy also sits on the floor nearby. The surrounding space includes a wall and a glass sliding door.",
"caption_ann": "The image captures a lively scene at an indoor pool with a <1:patterned tile floor>. A group of <6,7,8,9,10:five children> are playing in the <2:illuminated small pool>. A <6:girl with pigtails, wearing a black swimsuit with pink straps>, has her back to the camera and is holding onto the pool's edge, near a <3:curved metal handrail>. A <8:smiling boy> is in the center of the pool, and <7,9,10:three other children> are also visible in the water. In the background, a <16:pink and white patterned backpack> rests on the floor, and there is a seating area where several <11,12,13,14,15:five people> are gathered around a table on <17,18,19,20,21:five chairs>. A <4:purple and blue inflatable toy> also sits on the <1:floor> nearby. The surrounding space includes a <0:wall> and a <5:glass sliding door>.",
"id": 1769,
"image_id": "27_97-FeRYjxXU_00001472",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "patterned tile floor"
},
{
"mask_ids": [
6,
7,
8,
9,
10
],
"txt_desc": "five children"
},
{
"mask_ids": [
2
],
"txt_desc": "illuminated small pool"
},
{
"mask_ids": [
6
],
"txt_desc": "girl with pigtails, wearing a black swimsuit with pink straps"
},
{
"mask_ids": [
3
],
"txt_desc": "curved metal handrail"
},
{
"mask_ids": [
8
],
"txt_desc": "smiling boy"
},
{
"mask_ids": [
7,
9,
10
],
"txt_desc": "three other children"
},
{
"mask_ids": [
16
],
"txt_desc": "pink and white patterned backpack"
},
{
"mask_ids": [
11,
12,
13,
14,
15
],
"txt_desc": "five people"
},
{
"mask_ids": [
17,
18,
19,
20,
21
],
"txt_desc": "five chairs"
},
{
"mask_ids": [
4
],
"txt_desc": "purple and blue inflatable toy"
},
{
"mask_ids": [
1
],
"txt_desc": "floor"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
},
{
"mask_ids": [
5
],
"txt_desc": "glass sliding door"
}
],
"labels": [
"wall",
"floor",
"water",
"pipeline",
"toy",
"door",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"bag_or_package",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat"
]
} | [
{
"area": 55098,
"bbox": [
1,
0,
1278,
136
],
"category_id": 0,
"id": 19104,
"image_id": "27_97-FeRYjxXU_00001472",
"iscrowd": 0,
"segmentation": {
"counts": "Sh0k1cc0R1_Oa000000001O000000000000000000000000001O000000000000001O000000000000000000000000001O... | [
{
"id": 1,
"name": "object"
}
] |
train | 280_8ARSWpfl_9I_00001090 | 280_8ARSWpfl_9I_00001090.jpg | {
"data_source": "VIPSeg",
"file_name": "280_8ARSWpfl_9I_00001090.jpg",
"height": 720,
"id": "280_8ARSWpfl_9I_00001090",
"width": 1280
} | {
"caption": "In a busy outdoor scene, likely a college campus or city plaza, a group of young people are going about their day. In the foreground, a young man with curly hair, a beard, and a tattooed arm is engaged in an animated conversation, gesturing as he speaks and holding his cell phone. He is talking to a young woman with blonde hair and glasses, carrying a red backpack. Between them, a young woman in a striped shirt and a skirt walks by, looking at the man and holding her phone and a handbag . A man in a yellow tank top and a white cap rides his bicycle past the group. A mobile is affixed on his bicycle. Two persons and four bicycles are visible in the background. A person wearing black shirt and gray pants visible from the back is carrying a black bagpack. The scene is set on a paved brick plaza in front of a modern brick building, with trees in the background and bushes at the bottom of the building.",
"caption_ann": "In a busy outdoor scene, likely a college campus or city plaza, a group of young people are going about their day. In the foreground, a <14:young man with curly hair, a beard, and a tattooed arm> is engaged in an animated conversation, gesturing as he speaks and holding his <18:cell phone>. He is talking to a <10:young woman with blonde hair and glasses>, carrying a <16:red backpack>. Between them, a <9:young woman in a striped shirt and a skirt> walks by, looking at the man and holding her <20:phone> and a <15:handbag> . A <12:man in a yellow tank top and a white cap> rides his <5:bicycle> past the group. A <19:mobile> is affixed on his <5:bicycle>. <11,13:Two persons> and <4,6,7,8:four bicycles> are visible in the background. A <11:person wearing black shirt and gray pants> visible from the back is carrying a <17: black bagpack>. The scene is set on a <0:paved brick plaza> in front of a modern <1:brick building>, with <2:trees> in the background and <3:bushes> at the bottom of the building.",
"id": 1770,
"image_id": "280_8ARSWpfl_9I_00001090",
"label_matched": [
{
"mask_ids": [
14
],
"txt_desc": "young man with curly hair, a beard, and a tattooed arm"
},
{
"mask_ids": [
18
],
"txt_desc": "cell phone"
},
{
"mask_ids": [
10
],
"txt_desc": "young woman with blonde hair and glasses"
},
{
"mask_ids": [
16
],
"txt_desc": "red backpack"
},
{
"mask_ids": [
9
],
"txt_desc": "young woman in a striped shirt and a skirt"
},
{
"mask_ids": [
20
],
"txt_desc": "phone"
},
{
"mask_ids": [
15
],
"txt_desc": "handbag"
},
{
"mask_ids": [
12
],
"txt_desc": "man in a yellow tank top and a white cap"
},
{
"mask_ids": [
5
],
"txt_desc": "bicycle"
},
{
"mask_ids": [
19
],
"txt_desc": "mobile"
},
{
"mask_ids": [
5
],
"txt_desc": "bicycle"
},
{
"mask_ids": [
11,
13
],
"txt_desc": "Two persons"
},
{
"mask_ids": [
4,
6,
7,
8
],
"txt_desc": "four bicycles"
},
{
"mask_ids": [
11
],
"txt_desc": "person wearing black shirt and gray pants"
},
{
"mask_ids": [
17
],
"txt_desc": "black bagpack"
},
{
"mask_ids": [
0
],
"txt_desc": "paved brick plaza"
},
{
"mask_ids": [
1
],
"txt_desc": "brick building"
},
{
"mask_ids": [
2
],
"txt_desc": "trees"
},
{
"mask_ids": [
3
],
"txt_desc": "bushes"
}
],
"labels": [
"ground",
"building",
"tree",
"other_plant",
"bicycle",
"bicycle",
"bicycle",
"bicycle",
"bicycle",
"person",
"person",
"person",
"person",
"person",
"person",
"bag_or_package",
"bag_or_package",
"bag_or_package",
"Mobile_phone",
"Mobile_phone",
"Mobile_phone"
]
} | [
{
"area": 100812,
"bbox": [
0,
519,
1280,
201
],
"category_id": 14,
"id": 19126,
"image_id": "280_8ARSWpfl_9I_00001090",
"iscrowd": 0,
"segmentation": {
"counts": "Yf07Wf02N2M3N2M3N2M3N2M3N2N2000000O100O100000000004L9G5K00001O1O00001O001O00001O1O00001O00... | [
{
"id": 1,
"name": "object"
}
] |
train | 281_9FJZRZ5eXa4_00000790 | 281_9FJZRZ5eXa4_00000790.jpg | {
"data_source": "VIPSeg",
"file_name": "281_9FJZRZ5eXa4_00000790.jpg",
"height": 720,
"id": "281_9FJZRZ5eXa4_00000790",
"width": 1280
} | {
"caption": "The image features the front of a large, light-coloured building with a distinctive stone facade, under a cloudy light-grey sky. In front of the building, a large billboard for the Warner College of Natural Resources is visible. In the bottom right corner bicycle stand can be seen. In the foreground a man wearing a white t-shirt is playing a guitar next to a girl. In the bottom left corner two girls are talking, and a small leafless plant can be seen next to the billboard.",
"caption_ann": "The image features the front of a <0:large, light-coloured building> with a distinctive stone facade, under a cloudy <1:light-grey sky>. In front of the building, a <2:large billboard for the Warner College of Natural Resources> is visible. In the bottom right corner <4:bicycle stand> can be seen. In the foreground a <5:man wearing a white t-shirt> is playing a <9:guitar> next to a <6:girl>. In the bottom left corner <7,8:two girls> are talking, and a <3:small leafless plant> can be seen next to the <2:billboard>.",
"id": 1771,
"image_id": "281_9FJZRZ5eXa4_00000790",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "large, light-coloured building"
},
{
"mask_ids": [
1
],
"txt_desc": "light-grey sky"
},
{
"mask_ids": [
2
],
"txt_desc": "large billboard for the Warner College of Natural Resources"
},
{
"mask_ids": [
4
],
"txt_desc": "bicycle stand"
},
{
"mask_ids": [
5
],
"txt_desc": "man wearing a white t-shirt"
},
{
"mask_ids": [
9
],
"txt_desc": "guitar"
},
{
"mask_ids": [
6
],
"txt_desc": "girl"
},
{
"mask_ids": [
7,
8
],
"txt_desc": "two girls"
},
{
"mask_ids": [
3
],
"txt_desc": "small leafless plant"
},
{
"mask_ids": [
2
],
"txt_desc": "billboard"
}
],
"labels": [
"bridge",
"sky",
"billboard_or_Bulletin_Board",
"other_plant",
"bicycle",
"person",
"person",
"person",
"person",
"instrument"
]
} | [
{
"area": 687179,
"bbox": [
0,
0,
1280,
720
],
"category_id": 23,
"id": 19147,
"image_id": "281_9FJZRZ5eXa4_00000790",
"iscrowd": 0,
"segmentation": {
"counts": "0oc0a200O100O100O1N2O1N2N2O1N2O10000O10000O10000O1N2O1O1N2O1N2N2M3N2M3M3N20000O10000O1000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 282_BiPjVs6Dugs_00000145 | 282_BiPjVs6Dugs_00000145.jpg | {
"data_source": "VIPSeg",
"file_name": "282_BiPjVs6Dugs_00000145.jpg",
"height": 720,
"id": "282_BiPjVs6Dugs_00000145",
"width": 1280
} | {
"caption": "The image shows a concrete bridge with the word \"UNIVERSITY\" engraved. On the bridge, four people are standing behind a black metal railing. On the middle left of the image, a guy with a red hooded sweatshirt can be seen headed towards a woman in a red sweatshirt who is gesturing with her hands next to a woman in a grey sweatshirt. Further to the left corner of the image, a young woman with straight blonde hair is visible. The area surrounding the bridge features a large grassy area and bare tree.",
"caption_ann": "The image shows a <2:concrete bridge with the word \"UNIVERSITY\" engraved>. On the <2:bridge>, <4,5,6,7:four people> are standing behind a <0:black metal railing>. On the middle left of the image, a <5:guy with a red hooded sweatshirt> can be seen headed towards a <6:woman in a red sweatshirt> who is gesturing with her hands next to a <7:woman in a grey sweatshirt>. Further to the left corner of the image, a <4:young woman with straight blonde hair> is visible. The area surrounding the bridge features a <1:large grassy area> and <3:bare tree>.",
"id": 1772,
"image_id": "282_BiPjVs6Dugs_00000145",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "concrete bridge with the word \"UNIVERSITY\" engraved"
},
{
"mask_ids": [
2
],
"txt_desc": "bridge"
},
{
"mask_ids": [
4,
5,
6,
7
],
"txt_desc": "four people"
},
{
"mask_ids": [
0
],
"txt_desc": "black metal railing"
},
{
"mask_ids": [
5
],
"txt_desc": "guy with a red hooded sweatshirt"
},
{
"mask_ids": [
6
],
"txt_desc": "woman in a red sweatshirt"
},
{
"mask_ids": [
7
],
"txt_desc": "woman in a grey sweatshirt"
},
{
"mask_ids": [
4
],
"txt_desc": "young woman with straight blonde hair"
},
{
"mask_ids": [
1
],
"txt_desc": "large grassy area"
},
{
"mask_ids": [
3
],
"txt_desc": "bare tree"
}
],
"labels": [
"handrail_or_fence",
"grass",
"bridge",
"tree",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 97904,
"bbox": [
0,
334,
1280,
210
],
"category_id": 7,
"id": 19157,
"image_id": "282_BiPjVs6Dugs_00000145",
"iscrowd": 0,
"segmentation": {
"counts": "m>S2]d0000000000O010000000O01000000000O01000000000O010000000O01000000000O010000000O010000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 288_RXHqZh47-HI_00000835 | 288_RXHqZh47-HI_00000835.jpg | {
"data_source": "VIPSeg",
"file_name": "288_RXHqZh47-HI_00000835.jpg",
"height": 720,
"id": "288_RXHqZh47-HI_00000835",
"width": 1280
} | {
"caption": "Inside a bustling library, a person, seen from behind wearing a hooded sweatshirt carrying a large backpack walks through the aisle. Four students are seated at four separate wooden tables, focused on their work. Three of them are using their laptops. The tables are equipped with modern study lamps and various cups and bottles. The tables are surrounded by seven wooden chairs. Large windows line the room, offering a view of a leafy, outdoor campus scene.",
"caption_ann": "Inside a bustling library, a <4:person, seen from behind wearing a hooded sweatshirt> carrying a <7:large backpack> walks through the aisle. <2,3,5,6:Four students> are seated at <9,10,11,12:four separate wooden tables>, focused on their work. Three of them are using <20,21,22:their laptops>. The tables are equipped with modern <0:study lamps> and various <8:cups and bottles>. The tables are surrounded by <13,14,15,16,17,18,19: seven wooden chairs>. Large <1:windows> line the room, offering a view of a leafy, outdoor campus scene.",
"id": 1773,
"image_id": "288_RXHqZh47-HI_00000835",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "person, seen from behind wearing a hooded sweatshirt"
},
{
"mask_ids": [
7
],
"txt_desc": "large backpack"
},
{
"mask_ids": [
2,
3,
5,
6
],
"txt_desc": "Four students"
},
{
"mask_ids": [
9,
10,
11,
12
],
"txt_desc": "four separate wooden tables"
},
{
"mask_ids": [
20,
21,
22
],
"txt_desc": "their laptops"
},
{
"mask_ids": [
0
],
"txt_desc": "study lamps"
},
{
"mask_ids": [
8
],
"txt_desc": "cups and bottles"
},
{
"mask_ids": [
13,
14,
15,
16,
17,
18,
19
],
"txt_desc": "seven wooden chairs"
},
{
"mask_ids": [
1
],
"txt_desc": "windows"
}
],
"labels": [
"lamp",
"window",
"person",
"person",
"person",
"person",
"person",
"bag_or_package",
"bottle_or_cup",
"table_or_desk",
"table_or_desk",
"table_or_desk",
"table_or_desk",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"chair_or_seat",
"computer",
"computer",
"computer"
]
} | [
{
"area": 5712,
"bbox": [
1162,
174,
118,
57
],
"category_id": 59,
"id": 19165,
"image_id": "288_RXHqZh47-HI_00000835",
"iscrowd": 0,
"segmentation": {
"counts": "TXai03Zf03L4M3L4M3L4M3L4M3L4M3L4M3N2000000000000000O10O10000000000000000000000000000O100000... | [
{
"id": 1,
"name": "object"
}
] |
train | 288_lYx1-qDVYjU_00001052 | 288_lYx1-qDVYjU_00001052.jpg | {
"data_source": "VIPSeg",
"file_name": "288_lYx1-qDVYjU_00001052.jpg",
"height": 720,
"id": "288_lYx1-qDVYjU_00001052",
"width": 1280
} | {
"caption": "The image shows a large set of stone stairs forming the seating area, with a low stone wall and a grassy hill separating the seating from the central area. The central area is covered by a gravel-like ground. In the middle of the frame, a man wearing a grey and black outfit is walking away from the viewer. To his left, a man wearing a blue jacket is standing still. On the bottom right corner, a man wearing a dark blue jacket and a woman wearing a black jacket are headed towards the stone stairs. In the background, a row of large, green trees lines the top edge of the amphitheatre, under a bright blue sky.",
"caption_ann": "The image shows a <1:large set of stone stairs> forming the seating area, with a <0:low stone wall> and a <3:grassy hill> separating the seating from the central area. The central area is covered by a <2:gravel-like ground>. In the middle of the frame, a <7:man wearing a grey and black outfit> is walking away from the viewer. To his left, a <6:man wearing a blue jacket> is standing still. On the bottom right corner, a <9:man wearing a dark blue jacket> and a <8:woman wearing a black jacket> are headed towards the <1:stone stairs>. In the background, a <5:row of large, green trees> lines the top edge of the amphitheatre, under a bright <4:blue sky>.",
"id": 1774,
"image_id": "288_lYx1-qDVYjU_00001052",
"label_matched": [
{
"mask_ids": [
1
],
"txt_desc": "large set of stone stairs"
},
{
"mask_ids": [
0
],
"txt_desc": "low stone wall"
},
{
"mask_ids": [
3
],
"txt_desc": "grassy hill"
},
{
"mask_ids": [
2
],
"txt_desc": "gravel-like ground"
},
{
"mask_ids": [
7
],
"txt_desc": "man wearing a grey and black outfit"
},
{
"mask_ids": [
6
],
"txt_desc": "man wearing a blue jacket"
},
{
"mask_ids": [
9
],
"txt_desc": "man wearing a dark blue jacket"
},
{
"mask_ids": [
8
],
"txt_desc": "woman wearing a black jacket"
},
{
"mask_ids": [
1
],
"txt_desc": "stone stairs"
},
{
"mask_ids": [
5
],
"txt_desc": "row of large, green trees"
},
{
"mask_ids": [
4
],
"txt_desc": "blue sky"
}
],
"labels": [
"wall",
"stair",
"ground",
"grass",
"sky",
"tree",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 181476,
"bbox": [
0,
85,
1280,
380
],
"category_id": 0,
"id": 19188,
"image_id": "288_lYx1-qDVYjU_00001052",
"iscrowd": 0,
"segmentation": {
"counts": "e2R2Y5_2\\5`MdJ_2]5aMcJ_2]5aMcJ_2]5aMcJ_2]5aMcJ_2]5aMcJ_2]5aMcJ_2]5aMcJ_2]5aMcJ_2]5aMcJ_2^5`... | [
{
"id": 1,
"name": "object"
}
] |
train | 28_GVaT-gP39KA_00000190 | 28_GVaT-gP39KA_00000190.jpg | {
"data_source": "VIPSeg",
"file_name": "28_GVaT-gP39KA_00000190.jpg",
"height": 720,
"id": "28_GVaT-gP39KA_00000190",
"width": 1278
} | {
"caption": "This aerial view captures a vast, wide river during winter, with a large expanse of its surface filled with broken, white ice floes. In the lower right corner, a small, partially snow-covered ship is nestled within the ice. A channel of dark, open river water separates the ice pack from the distant shore, which is lined with a dense cityscape of numerous buildings and skyscrapers. Above the city, the scene is capped by a pale overcast sky.",
"caption_ann": "This aerial view captures a <3:vast, wide river> during winter, with a large expanse of its surface filled with <2:broken, white ice floes>. In the lower right corner, a <4:small, partially snow-covered ship> is nestled within the ice. A <3:channel of dark, open river water> separates the <2:ice pack> from the distant shore, which is lined with a dense <0:cityscape of numerous buildings and skyscrapers>. Above the city, the scene is capped by a <1:pale overcast sky>.",
"id": 1775,
"image_id": "28_GVaT-gP39KA_00000190",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "vast, wide river"
},
{
"mask_ids": [
2
],
"txt_desc": "broken, white ice floes"
},
{
"mask_ids": [
4
],
"txt_desc": "small, partially snow-covered ship"
},
{
"mask_ids": [
3
],
"txt_desc": "channel of dark, open river water"
},
{
"mask_ids": [
2
],
"txt_desc": "ice pack"
},
{
"mask_ids": [
0
],
"txt_desc": "cityscape of numerous buildings and skyscrapers"
},
{
"mask_ids": [
1
],
"txt_desc": "pale overcast sky"
}
],
"labels": [
"building",
"sky",
"ice",
"river",
"ship_or_boat"
]
} | [
{
"area": 135596,
"bbox": [
0,
34,
1278,
216
],
"category_id": 21,
"id": 19198,
"image_id": "28_GVaT-gP39KA_00000190",
"iscrowd": 0,
"segmentation": {
"counts": "f3T4\\b0O100000000000000000000000000N1100000000000000000000000O1000N200000000000000000N20000... | [
{
"id": 1,
"name": "object"
}
] |
train | 290_tpK57_NiSEs_00000497 | 290_tpK57_NiSEs_00000497.jpg | {
"data_source": "VIPSeg",
"file_name": "290_tpK57_NiSEs_00000497.jpg",
"height": 720,
"id": "290_tpK57_NiSEs_00000497",
"width": 1280
} | {
"caption": "A person in black outfit sits on the ancient stone steps of a large amphitheater, while another person is barely visible in the distance near an open grassy area and a line of green trees. Another ancient construction is visible behind the amphitheater against the sky in the background.",
"caption_ann": "A <7:person in black outfit> sits on the <0:ancient stone steps of a large amphitheater>, while another <5:person> is barely visible in the distance near an <1:open grassy area> and a <4:line of green trees>. <2:Another ancient construction> is visible behind the <0:amphitheater> against the <3:sky> in the background.",
"id": 1776,
"image_id": "290_tpK57_NiSEs_00000497",
"label_matched": [
{
"mask_ids": [
7
],
"txt_desc": "person in black outfit"
},
{
"mask_ids": [
0
],
"txt_desc": "ancient stone steps of a large amphitheater"
},
{
"mask_ids": [
5
],
"txt_desc": "person"
},
{
"mask_ids": [
1
],
"txt_desc": "open grassy area"
},
{
"mask_ids": [
4
],
"txt_desc": "line of green trees"
},
{
"mask_ids": [
2
],
"txt_desc": "Another ancient construction"
},
{
"mask_ids": [
0
],
"txt_desc": "amphitheater"
},
{
"mask_ids": [
3
],
"txt_desc": "sky"
}
],
"labels": [
"stair",
"ground",
"other_construction",
"sky",
"tree",
"person",
"person",
"person"
]
} | [
{
"area": 799436,
"bbox": [
0,
0,
1280,
720
],
"category_id": 3,
"id": 19203,
"image_id": "290_tpK57_NiSEs_00000497",
"iscrowd": 0,
"segmentation": {
"counts": "o7a>o70000000000000000000000000000000000000000000000000000000000000001O0000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 292_7Z2ZBLkZZUQ_00001760 | 292_7Z2ZBLkZZUQ_00001760.jpg | {
"data_source": "VIPSeg",
"file_name": "292_7Z2ZBLkZZUQ_00001760.jpg",
"height": 720,
"id": "292_7Z2ZBLkZZUQ_00001760",
"width": 1280
} | {
"caption": "This image displays a two young girls playing arcade racing games. The girl on the right in yellow pants is wearing a black backpack and sits behind the girl on the left, who is holding a pink stuffed toy while playing. Colorful arcade flooring and walls fill the background.",
"caption_ann": "This image displays a <4,5:two young girls> playing <3:arcade racing games>. The <4:girl on the right in yellow pants> is wearing a <6:black backpack> and sits behind the <5:girl on the left>, who is holding a <2:pink stuffed toy> while playing. Colorful <1:arcade flooring> and <0:walls> fill the background.",
"id": 1777,
"image_id": "292_7Z2ZBLkZZUQ_00001760",
"label_matched": [
{
"mask_ids": [
4,
5
],
"txt_desc": "two young girls"
},
{
"mask_ids": [
3
],
"txt_desc": "arcade racing games"
},
{
"mask_ids": [
4
],
"txt_desc": "girl on the right in yellow pants"
},
{
"mask_ids": [
6
],
"txt_desc": "black backpack"
},
{
"mask_ids": [
5
],
"txt_desc": "girl on the left"
},
{
"mask_ids": [
2
],
"txt_desc": "pink stuffed toy"
},
{
"mask_ids": [
1
],
"txt_desc": "arcade flooring"
},
{
"mask_ids": [
0
],
"txt_desc": "walls"
}
],
"labels": [
"wall",
"floor",
"toy",
"other_electronic_product",
"person",
"person",
"bag_or_package"
]
} | [
{
"area": 20542,
"bbox": [
573,
0,
392,
349
],
"category_id": 0,
"id": 19211,
"image_id": "292_7Z2ZBLkZZUQ_00001760",
"iscrowd": 0,
"segmentation": {
"counts": "oVc<6Xf04L4L2nNEm[O>hc0NU\\O3bc09X\\OJec0;X\\OG`c0h1M3F9XJfNmCfNaNb4m<`4J6L5G8BcNUDQIb;T7dDjH... | [
{
"id": 1,
"name": "object"
}
] |
train | 294_tub3vR-v7mc_00000377 | 294_tub3vR-v7mc_00000377.jpg | {
"data_source": "VIPSeg",
"file_name": "294_tub3vR-v7mc_00000377.jpg",
"height": 720,
"id": "294_tub3vR-v7mc_00000377",
"width": 1280
} | {
"caption": "The image shows a man wearing a white t-shirt and red shorts positioned in the centre, facing the multicolour waterslide. To the left, a girl in a black shorts is crouched down at the top of the slide while a girl wearing a striped top is laying on her stomach to the right side of the image. Another person is partially visible on the bottom right corner. The waterslide descends through a lush, green landscape of trees and grassy ground. On the top left corner, a house is visible. The scene is set under a cloudy sky, with sky blue water slides visible in the background.",
"caption_ann": "The image shows a <8:man wearing a white t-shirt and red shorts> positioned in the centre, facing the <0:multicolour waterslide>. To the left, a <6:girl in a black shorts> is crouched down at the top of the slide while a <9:girl wearing a striped top> is laying on her stomach to the right side of the image. Another <7:person> is partially visible on the bottom right corner. The waterslide descends through a lush, <5:green landscape of trees> and <1:grassy ground>. On the top left corner, a <2:house> is visible. The scene is set under a cloudy <3:sky>, with <4:sky blue water slides> visible in the background.",
"id": 1778,
"image_id": "294_tub3vR-v7mc_00000377",
"label_matched": [
{
"mask_ids": [
8
],
"txt_desc": "man wearing a white t-shirt and red shorts"
},
{
"mask_ids": [
0
],
"txt_desc": "multicolour waterslide"
},
{
"mask_ids": [
6
],
"txt_desc": "girl in a black shorts"
},
{
"mask_ids": [
9
],
"txt_desc": "girl wearing a striped top"
},
{
"mask_ids": [
7
],
"txt_desc": "person"
},
{
"mask_ids": [
5
],
"txt_desc": "green landscape of trees"
},
{
"mask_ids": [
1
],
"txt_desc": "grassy ground"
},
{
"mask_ids": [
2
],
"txt_desc": "house"
},
{
"mask_ids": [
3
],
"txt_desc": "sky"
},
{
"mask_ids": [
4
],
"txt_desc": "sky blue water slides"
}
],
"labels": [
"Playground_slide",
"ground",
"house",
"sky",
"water",
"tree",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 483763,
"bbox": [
0,
4,
1280,
716
],
"category_id": 6,
"id": 19218,
"image_id": "294_tub3vR-v7mc_00000377",
"iscrowd": 0,
"segmentation": {
"counts": "4\\f0c0A8H?A8H?A8H=C3M8H3M7I4L4L2N3M1O4L3M5K3M5K4L5K3M5K3M5K1O0000O100O10000O1O1000000O100O10... | [
{
"id": 1,
"name": "object"
}
] |
train | 296_dy90PoMcwLI_00000730 | 296_dy90PoMcwLI_00000730.jpg | {
"data_source": "VIPSeg",
"file_name": "296_dy90PoMcwLI_00000730.jpg",
"height": 720,
"id": "296_dy90PoMcwLI_00000730",
"width": 1280
} | {
"caption": "A group of thirteen students are relaxing and socializing on a wide set of outdoor concrete steps that function as amphitheater-style seating. Some are sitting in small groups, chatting, while others are focused on books or papers. One young man is sitting with a young woman, both looking at something together. The steps are set in a modern campus-like environment, with paved ground at the bottom and a large building and lush green bushes in the background. Several planters are placed along the upper level. Various bags and backpacks are scattered around the students. A lamppost and circular metallic poles are also visible in the background.",
"caption_ann": "A group of <6,7,8,9,10,11,12,13,14,15,16,17,18:thirteen students> are relaxing and socializing on a wide set of <0:outdoor concrete steps> that function as amphitheater-style seating. Some are sitting in small groups, chatting, while others are focused on books or papers. One <17:young man> is sitting with a <16:young woman>, both looking at something together. The steps are set in a modern campus-like environment, with <2:paved ground> at the bottom and a <3:large building> and <5:lush green bushes> in the background. Several <21,22,23,24:planters> are placed along the upper level. Various <19,20:bags and backpacks> are scattered around the students. A <4:lamppost> and <1:circular metallic poles> are also visible in the background.",
"id": 1779,
"image_id": "296_dy90PoMcwLI_00000730",
"label_matched": [
{
"mask_ids": [
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18
],
"txt_desc": "thirteen students"
},
{
"mask_ids": [
0
],
"txt_desc": "outdoor concrete steps"
},
{
"mask_ids": [
17
],
"txt_desc": "young man"
},
{
"mask_ids": [
16
],
"txt_desc": "young woman"
},
{
"mask_ids": [
2
],
"txt_desc": "paved ground"
},
{
"mask_ids": [
3
],
"txt_desc": "large building"
},
{
"mask_ids": [
5
],
"txt_desc": "lush green bushes"
},
{
"mask_ids": [
21,
22,
23,
24
],
"txt_desc": "planters"
},
{
"mask_ids": [
19,
20
],
"txt_desc": "bags and backpacks"
},
{
"mask_ids": [
4
],
"txt_desc": "lamppost"
},
{
"mask_ids": [
1
],
"txt_desc": "circular metallic poles"
}
],
"labels": [
"stair",
"pole",
"ground",
"building",
"lamp",
"other_plant",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"bag_or_package",
"bag_or_package",
"flower_pot_or_vase",
"flower_pot_or_vase",
"flower_pot_or_vase",
"flower_pot_or_vase"
]
} | [
{
"area": 406760,
"bbox": [
0,
250,
1280,
470
],
"category_id": 3,
"id": 19228,
"image_id": "296_dy90PoMcwLI_00000730",
"iscrowd": 0,
"segmentation": {
"counts": "V8_3Qc0U2kM00001O00000000001O00001O000000000010O1O00100O1O1O1iJT_Oh4l`0XKU_Og4k`0YKX_Oe4g`0... | [
{
"id": 1,
"name": "object"
}
] |
train | 297_hfOFzH9XuK4_00001877 | 297_hfOFzH9XuK4_00001877.jpg | {
"data_source": "VIPSeg",
"file_name": "297_hfOFzH9XuK4_00001877.jpg",
"height": 720,
"id": "297_hfOFzH9XuK4_00001877",
"width": 1280
} | {
"caption": "This image shows a room with a light-coloured wall and a white panelled door on the right. The man dressed in a white long-sleeved shirt, is perched on a small wooden stool and points into the large aquarium. His companion, a man wearing a colourful cap and sunglasses, stands with a large fishing net with a long handle. The aquarium rests on a sturdy wooden stand, in front of which is a chair with a brown-coloured back. The aquarium houses a variety of fish, including a prominent large, brownish-coloured fish, a darker, elongated fish near the top, a dark fish near the bottom, and a group of six small, silvery fish. To the far left, a portion of another aquarium and a black table are visible on the left side.",
"caption_ann": "This image shows a room with a <0:light-coloured wall> and a <2:white panelled door> on the right. The <3:man dressed in a white long-sleeved shirt>, is perched on a <19:small wooden stool> and points into the <14:large aquarium>. His companion, a <4:man wearing a colourful cap and sunglasses>, stands with a <1:large fishing net with a long handle>. The <14:aquarium> rests on a <17:sturdy wooden stand>, in front of which is a <18:chair with a brown-coloured back>. The <14:aquarium> houses a variety of fish, including a prominent <5:large, brownish-coloured fish>, a <10:darker, elongated fish> near the top, a <9:dark fish> near the bottom, and a group of <6,7,8,11,12,13:six small, silvery fish>. To the far left, a portion of another <15:aquarium> and a <16:black table> are visible on the left side.",
"id": 1780,
"image_id": "297_hfOFzH9XuK4_00001877",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "light-coloured wall"
},
{
"mask_ids": [
2
],
"txt_desc": "white panelled door"
},
{
"mask_ids": [
3
],
"txt_desc": "man dressed in a white long-sleeved shirt"
},
{
"mask_ids": [
19
],
"txt_desc": "small wooden stool"
},
{
"mask_ids": [
14
],
"txt_desc": "large aquarium"
},
{
"mask_ids": [
4
],
"txt_desc": "man wearing a colourful cap and sunglasses"
},
{
"mask_ids": [
1
],
"txt_desc": "large fishing net with a long handle"
},
{
"mask_ids": [
14
],
"txt_desc": "aquarium"
},
{
"mask_ids": [
17
],
"txt_desc": "sturdy wooden stand"
},
{
"mask_ids": [
18
],
"txt_desc": "chair with a brown-coloured back"
},
{
"mask_ids": [
14
],
"txt_desc": "aquarium"
},
{
"mask_ids": [
5
],
"txt_desc": "large, brownish-coloured fish"
},
{
"mask_ids": [
10
],
"txt_desc": "darker, elongated fish"
},
{
"mask_ids": [
9
],
"txt_desc": "dark fish"
},
{
"mask_ids": [
6,
7,
8,
11,
12,
13
],
"txt_desc": "six small, silvery fish"
},
{
"mask_ids": [
15
],
"txt_desc": "aquarium"
},
{
"mask_ids": [
16
],
"txt_desc": "black table"
}
],
"labels": [
"wall",
"pole",
"door",
"person",
"person",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"fishbowl",
"fishbowl",
"table_or_desk",
"table_or_desk",
"chair_or_seat",
"chair_or_seat"
]
} | [
{
"area": 101097,
"bbox": [
1,
0,
1186,
720
],
"category_id": 0,
"id": 19253,
"image_id": "297_hfOFzH9XuK4_00001877",
"iscrowd": 0,
"segmentation": {
"counts": "`f0W1Ze0O00000001O000000001O000000000000000000001O000000001O000000000000000000001O00000000001... | [
{
"id": 1,
"name": "object"
}
] |
train | 298_emBoDloCze8_00002860 | 298_emBoDloCze8_00002860.jpg | {
"data_source": "VIPSeg",
"file_name": "298_emBoDloCze8_00002860.jpg",
"height": 720,
"id": "298_emBoDloCze8_00002860",
"width": 1280
} | {
"caption": "A young man in a maroon hoodie and black shorts is standing on the edge of a concrete surface next to the tiled walkway, bending over to look in a large basin while holding a pair of white shoes in his right hand and is wearing a grey backpack. In the background, to the left, two boys are sitting on a part of a concrete ledge with a stair. A black fence is visible behind them, with a yellow excavator and a house in the far distance. A large, dark sculpture can be seen on the right side of the image, with water cascading over it. The entire scene is surrounded by numerous dense green trees.",
"caption_ann": "A <8:young man in a maroon hoodie and black shorts> is standing on the edge of a concrete surface next to the <2:tiled walkway>, bending over to look in a large basin while holding a pair of white shoes in his right hand and is wearing a <11:grey backpack>. In the background, to the left, <9,10:two boys> are sitting on a part of a <0:concrete ledge with a stair>. A <1:black fence> is visible behind them, with a <5:yellow excavator> and a <3:house> in the far distance. A large, dark <7:sculpture> can be seen on the right side of the image, with <4:water> cascading over it. The entire scene is surrounded by numerous <6:dense green trees>.",
"id": 1781,
"image_id": "298_emBoDloCze8_00002860",
"label_matched": [
{
"mask_ids": [
8
],
"txt_desc": "young man in a maroon hoodie and black shorts"
},
{
"mask_ids": [
2
],
"txt_desc": "tiled walkway"
},
{
"mask_ids": [
11
],
"txt_desc": "grey backpack"
},
{
"mask_ids": [
9,
10
],
"txt_desc": "two boys"
},
{
"mask_ids": [
0
],
"txt_desc": "concrete ledge with a stair"
},
{
"mask_ids": [
1
],
"txt_desc": "black fence"
},
{
"mask_ids": [
5
],
"txt_desc": "yellow excavator"
},
{
"mask_ids": [
3
],
"txt_desc": "house"
},
{
"mask_ids": [
7
],
"txt_desc": "sculpture"
},
{
"mask_ids": [
4
],
"txt_desc": "water"
},
{
"mask_ids": [
6
],
"txt_desc": "dense green trees"
}
],
"labels": [
"stair",
"handrail_or_fence",
"ground",
"house",
"water",
"wheeled_machine",
"tree",
"sculpture",
"person",
"person",
"person",
"bag_or_package"
]
} | [
{
"area": 50009,
"bbox": [
0,
280,
928,
121
],
"category_id": 3,
"id": 19273,
"image_id": "298_emBoDloCze8_00002860",
"iscrowd": 0,
"segmentation": {
"counts": "e9l2dc00000000000000000000000O1000000000000000000000000000000000000000O100000000000O100O1O1O1... | [
{
"id": 1,
"name": "object"
}
] |
train | 298_ma7hngzCd-I_00001337 | 298_ma7hngzCd-I_00001337.jpg | {
"data_source": "VIPSeg",
"file_name": "298_ma7hngzCd-I_00001337.jpg",
"height": 720,
"id": "298_ma7hngzCd-I_00001337",
"width": 1280
} | {
"caption": "The image displays a small child in a colorful outfit, with their right hand touching their chin as they look at a large aquarium placed on the floor which contains ten small sharks swimming in the water. The water is reflecting the light from the ceiling, creating a ripple effect on the surface.",
"caption_ann": "The image displays a small <2:child in a colorful outfit>, with their right hand touching their chin as they look at a large <13:aquarium> placed on the <1:floor> which contains <3,4,5,6,7,8,9,10,11,12:ten small sharks> swimming in the water. The water is reflecting the light from the <0:ceiling>, creating a ripple effect on the surface.",
"id": 1782,
"image_id": "298_ma7hngzCd-I_00001337",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "child in a colorful outfit"
},
{
"mask_ids": [
13
],
"txt_desc": "aquarium"
},
{
"mask_ids": [
1
],
"txt_desc": "floor"
},
{
"mask_ids": [
3,
4,
5,
6,
7,
8,
9,
10,
11,
12
],
"txt_desc": "ten small sharks"
},
{
"mask_ids": [
0
],
"txt_desc": "ceiling"
}
],
"labels": [
"ceiling",
"floor",
"person",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"other_animal",
"fishbowl"
]
} | [
{
"area": 18765,
"bbox": [
588,
0,
420,
102
],
"category_id": 1,
"id": 19285,
"image_id": "298_ma7hngzCd-I_00001337",
"iscrowd": 0,
"segmentation": {
"counts": "P^m<1_f0000000001O00001O00000000001O00001O000000001O00001O00000000001O001O00000000001O001O000... | [
{
"id": 1,
"name": "object"
}
] |
train | 299_CUjnxsP1fIM_00001097 | 299_CUjnxsP1fIM_00001097.jpg | {
"data_source": "VIPSeg",
"file_name": "299_CUjnxsP1fIM_00001097.jpg",
"height": 720,
"id": "299_CUjnxsP1fIM_00001097",
"width": 1280
} | {
"caption": "The image features an outdoor site that is covered in a large number of rocks and dirt ground. Three people are working at the site: the man on the left wearing a light grey shirt is kneeling and appears to be sifting through the rocky area, the person wearing a purple top is walking in the back, and a person in a white t-shirt and a hat is walking toward a wheelbarrow. Several barrels are in use and scattered around the area. In the background, there is a dense green bushes just above the rocky terrain with a clear, light blue sky visible above it.",
"caption_ann": "The image features an outdoor site that is covered in a large number of <2:rocks> and <0:dirt ground>. <5,6,7:Three people> are working at the site: the <7:man on the left wearing a light grey shirt> is kneeling and appears to be sifting through the <2:rocky area>, the <6:person wearing a purple top> is walking in the back, and a <5:person in a white t-shirt and a hat> is walking toward a <3:wheelbarrow>. Several <8,9,10,11,12:barrels> are in use and scattered around the area. In the background, there is a <4:dense green bushes> just above the <2:rocky terrain> with a clear, <1:light blue sky> visible above it.",
"id": 1783,
"image_id": "299_CUjnxsP1fIM_00001097",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "rocks"
},
{
"mask_ids": [
0
],
"txt_desc": "dirt ground"
},
{
"mask_ids": [
5,
6,
7
],
"txt_desc": "Three people"
},
{
"mask_ids": [
7
],
"txt_desc": "man on the left wearing a light grey shirt"
},
{
"mask_ids": [
2
],
"txt_desc": "rocky area"
},
{
"mask_ids": [
6
],
"txt_desc": "person wearing a purple top"
},
{
"mask_ids": [
5
],
"txt_desc": "person in a white t-shirt and a hat"
},
{
"mask_ids": [
3
],
"txt_desc": "wheelbarrow"
},
{
"mask_ids": [
8,
9,
10,
11,
12
],
"txt_desc": "barrels"
},
{
"mask_ids": [
4
],
"txt_desc": "dense green bushes"
},
{
"mask_ids": [
2
],
"txt_desc": "rocky terrain"
},
{
"mask_ids": [
1
],
"txt_desc": "light blue sky"
}
],
"labels": [
"ground",
"sky",
"stone",
"wheeled_machine",
"other_plant",
"person",
"person",
"person",
"barrel",
"barrel",
"barrel",
"barrel",
"barrel"
]
} | [
{
"area": 180701,
"bbox": [
0,
252,
1280,
374
],
"category_id": 14,
"id": 19299,
"image_id": "299_CUjnxsP1fIM_00001097",
"iscrowd": 0,
"segmentation": {
"counts": "Z8R4]b001N2O1O1O1N2O1O0O2O1O1O1N2O1O1O1N101O1N2O1O11O000O2O00001O0000001O0O101O00001O00001... | [
{
"id": 1,
"name": "object"
}
] |
train | 299_gOmiwd2HS_c_00000940 | 299_gOmiwd2HS_c_00000940.jpg | {
"data_source": "VIPSeg",
"file_name": "299_gOmiwd2HS_c_00000940.jpg",
"height": 720,
"id": "299_gOmiwd2HS_c_00000940",
"width": 1281
} | {
"caption": "A group of people are enjoying a lovely day out for a picnic in a park. A man in a white t-shirt and sunglasses gives a shoulder ride to a small child wearing glasses, who is reaching up to touch the blossoms of a flowering tree that frames the shot. They are standing next to a colorful picnic blanket laid out on the field of green grass, where a woman in blue jacket, a woman wearing dark colored jacket and a man wearing olive jacket, blue jeans and a hat are sitting and relaxing. The blanket is laden with four picnic bags. Two teenagers are standing in the park with their backside facing the camera. A person in long black jacket, standing on the left side of the image is partially obscured by the flowers of the tree. A person wearing gray shirt and blue jeans has long hair and standing to the far right of the image. In the background, paved walkway is visible across the field of the park.",
"caption_ann": "A group of people are enjoying a lovely day out for a picnic in a park. A <9:man in a white t-shirt and sunglasses> gives a shoulder ride to a <7:small child wearing glasses>, who is reaching up to touch the blossoms of a <3:flowering tree> that frames the shot. They are standing next to a <2:colorful picnic blanket> laid out on the <0:field of green grass>, where a <6:woman in blue jacket>, a <12:woman wearing dark colored jacket> and a <8:man wearing olive jacket, blue jeans and a hat> are sitting and relaxing. The blanket is laden with <13,14,15,16:four picnic bags >. <5,11:Two teenagers> are standing in the park with their backside facing the camera. A <10:person in long black jacket>, standing on the left side of the image is partially obscured by the flowers of the <3:tree>. A <4:person wearing gray shirt and blue jeans has long hair> and standing to the far right of the image. In the background, <1:paved walkway> is visible across the <0:field> of the park.",
"id": 1784,
"image_id": "299_gOmiwd2HS_c_00000940",
"label_matched": [
{
"mask_ids": [
9
],
"txt_desc": "man in a white t-shirt and sunglasses"
},
{
"mask_ids": [
7
],
"txt_desc": "small child wearing glasses"
},
{
"mask_ids": [
3
],
"txt_desc": "flowering tree"
},
{
"mask_ids": [
2
],
"txt_desc": "colorful picnic blanket"
},
{
"mask_ids": [
0
],
"txt_desc": "field of green grass"
},
{
"mask_ids": [
6
],
"txt_desc": "woman in blue jacket"
},
{
"mask_ids": [
12
],
"txt_desc": "woman wearing dark colored jacket"
},
{
"mask_ids": [
8
],
"txt_desc": "man wearing olive jacket, blue jeans and a hat"
},
{
"mask_ids": [
13,
14,
15,
16
],
"txt_desc": "four picnic bags"
},
{
"mask_ids": [
5,
11
],
"txt_desc": "Two teenagers"
},
{
"mask_ids": [
10
],
"txt_desc": "person in long black jacket"
},
{
"mask_ids": [
3
],
"txt_desc": "tree"
},
{
"mask_ids": [
4
],
"txt_desc": "person wearing gray shirt and blue jeans has long hair"
},
{
"mask_ids": [
1
],
"txt_desc": "paved walkway"
},
{
"mask_ids": [
0
],
"txt_desc": "field"
}
],
"labels": [
"grass",
"path",
"cushion_or_carpet",
"flower",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"bag_or_package",
"bag_or_package",
"bag_or_package",
"bag_or_package"
]
} | [
{
"area": 202910,
"bbox": [
0,
141,
1281,
579
],
"category_id": 15,
"id": 19312,
"image_id": "299_gOmiwd2HS_c_00000940",
"iscrowd": 0,
"segmentation": {
"counts": "[>9`1S6i<oIUCS6k<mIUCS6k<mIUCS6k<mIUCS6k<mIUCS6k<mIUCS6k<mIUCS6k<mIUCS6k<mIUCS6k<mIUCS6k<m... | [
{
"id": 1,
"name": "object"
}
] |
train | 311_-X7okpS9Ufc_00003385 | 311_-X7okpS9Ufc_00003385.jpg | {
"data_source": "VIPSeg",
"file_name": "311_-X7okpS9Ufc_00003385.jpg",
"height": 720,
"id": "311_-X7okpS9Ufc_00003385",
"width": 1280
} | {
"caption": "In a brightly lit classroom with lime green and pale white walls and a brown tiled floor, a teacher dressed in a red and cream salwar kameez conducts a lesson for eight young students in yellow uniforms while all are seated in rows at four light brown wooden desks paired with four benches, watching the teacher as she holds up a flashcard. The room features a wooden slatted shelf and a storage unit on the left, and an alphabet poster in the upper right corner of the image. Natural light streams in through two windows. The larger window is covered by a translucent white curtain. In the middle of the image, a collection of colourful children's books is neatly arranged on a small wooden table. Several notebooks are placed on a wooden bench at the corner behind the students.",
"caption_ann": "In a brightly lit classroom with <0:lime green and pale white walls> and a <1:brown tiled floor>, a <8:teacher dressed in a red and cream salwar kameez> conducts a lesson for <9,10,11,12,13,14,15,16:eight young students in yellow uniforms> while all are seated in rows at <17,18,20,21:four light brown wooden desks> paired with <22,23,24,25,2:four benches>, watching the <8:teacher> as she holds up a flashcard. The room features a <3:wooden slatted shelf> and a <2:storage unit> on the left, and an <27:alphabet poster> in the upper right corner of the image. Natural light streams in through <6,7:two windows>. The <7:larger window > is covered by a <4:translucent white curtain>. In the middle of the image, a collection of colourful <5:children's books> is neatly arranged on a <19:small wooden table>. Several notebooks are placed on a <26:wooden bench> at the corner behind the students.",
"id": 1785,
"image_id": "311_-X7okpS9Ufc_00003385",
"label_matched": [
{
"mask_ids": [
0
],
"txt_desc": "lime green and pale white walls"
},
{
"mask_ids": [
1
],
"txt_desc": "brown tiled floor"
},
{
"mask_ids": [
8
],
"txt_desc": "teacher dressed in a red and cream salwar kameez"
},
{
"mask_ids": [
9,
10,
11,
12,
13,
14,
15,
16
],
"txt_desc": "eight young students in yellow uniforms"
},
{
"mask_ids": [
17,
18,
20,
21
],
"txt_desc": "four light brown wooden desks"
},
{
"mask_ids": [
22,
23,
24,
25,
2
],
"txt_desc": "four benches"
},
{
"mask_ids": [
8
],
"txt_desc": "teacher"
},
{
"mask_ids": [
3
],
"txt_desc": "wooden slatted shelf"
},
{
"mask_ids": [
2
],
"txt_desc": "storage unit"
},
{
"mask_ids": [
27
],
"txt_desc": "alphabet poster"
},
{
"mask_ids": [
6,
7
],
"txt_desc": "two windows"
},
{
"mask_ids": [
7
],
"txt_desc": "larger window"
},
{
"mask_ids": [
4
],
"txt_desc": "translucent white curtain"
},
{
"mask_ids": [
5
],
"txt_desc": "children's books"
},
{
"mask_ids": [
19
],
"txt_desc": "small wooden table"
},
{
"mask_ids": [
26
],
"txt_desc": "wooden bench"
}
],
"labels": [
"wall",
"floor",
"cupboard_or_showcase_or_storage_rack",
"shelf",
"curtain",
"book",
"window",
"window",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"person",
"table_or_desk",
"table_or_desk",
"table_or_desk",
"table_or_desk",
"table_or_desk",
"bench",
"bench",
"bench",
"bench",
"bench",
"painting_or_poster"
]
} | [
{
"area": 122946,
"bbox": [
0,
0,
1277,
429
],
"category_id": 0,
"id": 19329,
"image_id": "311_-X7okpS9Ufc_00003385",
"iscrowd": 0,
"segmentation": {
"counts": "W7R4]b0100O1000000O100O1000000O10000O1000000O100O100000000O10O02O001O0O2O0O2O1O000O2O1O0O2O1O... | [
{
"id": 1,
"name": "object"
}
] |
train | 313_BC49-YwZ_r8_00000617 | 313_BC49-YwZ_r8_00000617.jpg | {
"data_source": "VIPSeg",
"file_name": "313_BC49-YwZ_r8_00000617.jpg",
"height": 720,
"id": "313_BC49-YwZ_r8_00000617",
"width": 1280
} | {
"caption": "A person in a blue shirt is styling the long black hair of another person wearing a red cape. In front of them, there is a small wooden table with a red stool and red and white large table or desk with a pink patterned curtain hanging to the side. The room features a light wall with a colored floral design and a white floor. ",
"caption_ann": "A <3:person in a blue shirt> is styling the <4:long black hair> of another <4:person wearing a red cape>. In front of them, there is a <6:small wooden table> with a <7:red stool> and <5:red and white large table or desk> with a <2:pink patterned curtain> hanging to the side. The room features a <0:light wall with a colored floral design> and a <1:white floor>. ",
"id": 1786,
"image_id": "313_BC49-YwZ_r8_00000617",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "person in a blue shirt"
},
{
"mask_ids": [
4
],
"txt_desc": "long black hair"
},
{
"mask_ids": [
4
],
"txt_desc": "person wearing a red cape"
},
{
"mask_ids": [
6
],
"txt_desc": "small wooden table"
},
{
"mask_ids": [
7
],
"txt_desc": "red stool"
},
{
"mask_ids": [
5
],
"txt_desc": "red and white large table or desk"
},
{
"mask_ids": [
2
],
"txt_desc": "pink patterned curtain"
},
{
"mask_ids": [
0
],
"txt_desc": "light wall with a colored floral design"
},
{
"mask_ids": [
1
],
"txt_desc": "white floor"
}
],
"labels": [
"wall",
"floor",
"curtain",
"person",
"person",
"table_or_desk",
"table_or_desk",
"chair_or_seat"
]
} | [
{
"area": 198850,
"bbox": [
161,
0,
1119,
508
],
"category_id": 0,
"id": 19357,
"image_id": "313_BC49-YwZ_r8_00000617",
"iscrowd": 0,
"segmentation": {
"counts": "o\\a32]f02N3N1O2N1N4L103M1N3M2O2N1N3N2M3N0O2O1N2O0O2O1N2O002M1O2O1O1N101N2O1O0O4L2O3M1N4L2O... | [
{
"id": 1,
"name": "object"
}
] |
train | 314_i_7xU-M28IM_00001292 | 314_i_7xU-M28IM_00001292.jpg | {
"data_source": "VIPSeg",
"file_name": "314_i_7xU-M28IM_00001292.jpg",
"height": 720,
"id": "314_i_7xU-M28IM_00001292",
"width": 1280
} | {
"caption": "A woman in a green top and a black apron is giving a facial to a woman in pink and green dress who is seated in a black reclining chair. A floral curtain hangs on the wall behind them and a glass cupboard is visible to the left. A white table is in the bottom left corner, matching the white square tile floor.",
"caption_ann": "A <5:woman in a green top and a black apron> is giving a facial to a <4:woman in pink and green dress> who is seated in a <7:black reclining chair>. A <3:floral curtain> hangs on the <0:wall> behind them and a <2:glass cupboard> is visible to the left. A <6:white table> is in the bottom left corner, matching the <1:white square tile floor>.",
"id": 1787,
"image_id": "314_i_7xU-M28IM_00001292",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "woman in a green top and a black apron"
},
{
"mask_ids": [
4
],
"txt_desc": "woman in pink and green dress"
},
{
"mask_ids": [
7
],
"txt_desc": "black reclining chair"
},
{
"mask_ids": [
3
],
"txt_desc": "floral curtain"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
},
{
"mask_ids": [
2
],
"txt_desc": "glass cupboard"
},
{
"mask_ids": [
6
],
"txt_desc": "white table"
},
{
"mask_ids": [
1
],
"txt_desc": "white square tile floor"
}
],
"labels": [
"wall",
"floor",
"cupboard_or_showcase_or_storage_rack",
"curtain",
"person",
"person",
"table_or_desk",
"chair_or_seat"
]
} | [
{
"area": 166998,
"bbox": [
0,
227,
475,
480
],
"category_id": 0,
"id": 19365,
"image_id": "314_i_7xU-M28IM_00001292",
"iscrowd": 0,
"segmentation": {
"counts": "S7R7f?H5K9G4L9G4L9G4L9G5K8H5K9G2N000000O10000O1000001O1O0O101O001N10005K3M6J2M7J3M5K3L4M1O3M... | [
{
"id": 1,
"name": "object"
}
] |
train | 315_6NmrMU0Eg10_00000992 | 315_6NmrMU0Eg10_00000992.jpg | {
"data_source": "VIPSeg",
"file_name": "315_6NmrMU0Eg10_00000992.jpg",
"height": 720,
"id": "315_6NmrMU0Eg10_00000992",
"width": 1280
} | {
"caption": "The image displays a woman in black tank top and another woman wearing black colored full sleeved t-shirt arranging the bedding on a large bed in a tastefully decorated bedroom. The bed has a dark wood headboard. two comfortable-looking, black colored armchairs are placed at the front end of the bed. The armchair on the left is holding a white throw pillow on its top, and the armchair on the right is holding an assortment of three throw pillows. The bedroom features a neutral-toned wall with a large, scenic landscape painting hanging above the bed and a white ceiling. A pair of matching nightstands with a stylish lamp sits beside the bed on both sides. In the background, on the right, a large mirrored closet door reflects the room, and the floor appears to be carpeted.",
"caption_ann": "The image displays a <4:woman in black tank top> and another <5:woman wearing black colored full sleeved t-shirt> arranging the bedding on a <6:large bed> in a tastefully decorated bedroom. The <6:bed> has a dark wood headboard. <12,13:two comfortable-looking, black colored armchairs> are placed at the front end of the <6:bed>. The <12:armchair on the left> is holding a <10:white throw pillow> on its top, and the <13:armchair on the right> is holding an <7,8,9:assortment of three throw pillows>. The bedroom features a <0:neutral-toned wall> with a <14:large, scenic landscape painting> hanging above the bed and a <1:white ceiling>. A pair of matching <11:nightstands> with a stylish <3:lamp> sits beside the bed on both sides. In the background, on the right, a large <15:mirrored closet door> reflects the room, and the <2:floor> appears to be carpeted.",
"id": 1788,
"image_id": "315_6NmrMU0Eg10_00000992",
"label_matched": [
{
"mask_ids": [
4
],
"txt_desc": "woman in black tank top"
},
{
"mask_ids": [
5
],
"txt_desc": "woman wearing black colored full sleeved t-shirt"
},
{
"mask_ids": [
6
],
"txt_desc": "large bed"
},
{
"mask_ids": [
6
],
"txt_desc": "bed"
},
{
"mask_ids": [
12,
13
],
"txt_desc": "two comfortable-looking, black colored armchairs"
},
{
"mask_ids": [
6
],
"txt_desc": "bed"
},
{
"mask_ids": [
12
],
"txt_desc": "armchair on the left"
},
{
"mask_ids": [
10
],
"txt_desc": "white throw pillow"
},
{
"mask_ids": [
13
],
"txt_desc": "armchair on the right"
},
{
"mask_ids": [
7,
8,
9
],
"txt_desc": "assortment of three throw pillows"
},
{
"mask_ids": [
0
],
"txt_desc": "neutral-toned wall"
},
{
"mask_ids": [
14
],
"txt_desc": "large, scenic landscape painting"
},
{
"mask_ids": [
1
],
"txt_desc": "white ceiling"
},
{
"mask_ids": [
11
],
"txt_desc": "nightstands"
},
{
"mask_ids": [
3
],
"txt_desc": "lamp"
},
{
"mask_ids": [
15
],
"txt_desc": "mirrored closet door"
},
{
"mask_ids": [
2
],
"txt_desc": "floor"
}
],
"labels": [
"wall",
"ceiling",
"floor",
"lamp",
"person",
"person",
"bed",
"pillow",
"pillow",
"pillow",
"pillow",
"table_or_desk",
"chair_or_seat",
"chair_or_seat",
"painting_or_poster",
"mirror"
]
} | [
{
"area": 330736,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 19373,
"image_id": "315_6NmrMU0Eg10_00000992",
"iscrowd": 0,
"segmentation": {
"counts": "0oc0X1Y1Y1gN000000000O10000O1000000000000O10000O1000000000000O10000O10000N2N2J6L4J6N2J6L4J6N2... | [
{
"id": 1,
"name": "object"
}
] |
train | 318_h4eGtoua4y4_00001472 | 318_h4eGtoua4y4_00001472.jpg | {
"data_source": "VIPSeg",
"file_name": "318_h4eGtoua4y4_00001472.jpg",
"height": 720,
"id": "318_h4eGtoua4y4_00001472",
"width": 1280
} | {
"caption": "A father, wearing a baseball cap and a black shirt, helps his young son in a camouflage jacket to bowl. The child is using a green bowling ramp to aim a blue and black bowling ball down one of the polished wooden lanes of the bowling alley. In the background, the pins are set up, ready to be knocked down, and the alley is decorated with colorful wall graphics and advertisements. Another orange bowling ball is visible on an adjacent lane towards the far right side.",
"caption_ann": "A <3:father, wearing a baseball cap and a black shirt,> helps his <4:young son in a camouflage jacket> to bowl. The <4:child> is using a <2:green bowling ramp> to aim a <5:blue and black bowling ball> down one of the polished <1:wooden lanes> of the bowling alley. In the background, the pins are set up, ready to be knocked down, and the alley is decorated with colorful <0:wall graphics and advertisements>. Another <6:orange bowling ball> is visible on an adjacent lane towards the far right side.",
"id": 1789,
"image_id": "318_h4eGtoua4y4_00001472",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "father, wearing a baseball cap and a black shirt,"
},
{
"mask_ids": [
4
],
"txt_desc": "young son in a camouflage jacket"
},
{
"mask_ids": [
4
],
"txt_desc": "child"
},
{
"mask_ids": [
2
],
"txt_desc": "green bowling ramp"
},
{
"mask_ids": [
5
],
"txt_desc": "blue and black bowling ball"
},
{
"mask_ids": [
1
],
"txt_desc": "wooden lanes"
},
{
"mask_ids": [
0
],
"txt_desc": "wall graphics and advertisements"
},
{
"mask_ids": [
6
],
"txt_desc": "orange bowling ball"
}
],
"labels": [
"wall",
"floor",
"toy",
"person",
"person",
"ball",
"ball"
]
} | [
{
"area": 161489,
"bbox": [
0,
0,
1280,
159
],
"category_id": 0,
"id": 19389,
"image_id": "318_h4eGtoua4y4_00001472",
"iscrowd": 0,
"segmentation": {
"counts": "0Q4`b0O00000000000000000000000000000000000000000000000000000000000000000000000000000000000001... | [
{
"id": 1,
"name": "object"
}
] |
train | 319_l1Dz12fxQzQ_00000302 | 319_l1Dz12fxQzQ_00000302.jpg | {
"data_source": "VIPSeg",
"file_name": "319_l1Dz12fxQzQ_00000302.jpg",
"height": 720,
"id": "319_l1Dz12fxQzQ_00000302",
"width": 1280
} | {
"caption": "Two young boys are enjoying a game of bowling. In the foreground, a boy in a striped shirt and plaid shorts stands near the ball return machine, watching the lane. A pink bowling ball rests on the return. To the left, another boy in a red shirt holds a bowling ball as he prepares to take his turn. The bowling alley features a light-colored wooden floor and a large scoring monitor above the lanes, all set against a dark smooth wall. Multiple set of white colored bowling pins are positioned at the end of each bowling lane.",
"caption_ann": "Two <4,5:young boys> are enjoying a game of bowling. In the foreground, a <4:boy in a striped shirt and plaid shorts> stands near the <3:ball return machine>, watching the lane. A <6:pink bowling ball> rests on the return. To the left, another <5:boy in a red shirt> holds a <7:bowling ball> as he prepares to take his turn. The bowling alley features a <1:light-colored wooden floor> and a large <8:scoring monitor> above the lanes, all set against a <0:dark smooth wall>. Multiple <2:set of white colored bowling pins> are positioned at the end of each bowling lane.",
"id": 1790,
"image_id": "319_l1Dz12fxQzQ_00000302",
"label_matched": [
{
"mask_ids": [
4,
5
],
"txt_desc": "young boys"
},
{
"mask_ids": [
4
],
"txt_desc": "boy in a striped shirt and plaid shorts"
},
{
"mask_ids": [
3
],
"txt_desc": "ball return machine"
},
{
"mask_ids": [
6
],
"txt_desc": "pink bowling ball"
},
{
"mask_ids": [
5
],
"txt_desc": "boy in a red shirt"
},
{
"mask_ids": [
7
],
"txt_desc": "bowling ball"
},
{
"mask_ids": [
1
],
"txt_desc": "light-colored wooden floor"
},
{
"mask_ids": [
8
],
"txt_desc": "scoring monitor"
},
{
"mask_ids": [
0
],
"txt_desc": "dark smooth wall"
},
{
"mask_ids": [
2
],
"txt_desc": "set of white colored bowling pins"
}
],
"labels": [
"wall",
"floor",
"toy",
"other_machine",
"person",
"person",
"ball",
"ball",
"screen_or_television"
]
} | [
{
"area": 57701,
"bbox": [
0,
0,
1280,
219
],
"category_id": 0,
"id": 19396,
"image_id": "319_l1Dz12fxQzQ_00000302",
"iscrowd": 0,
"segmentation": {
"counts": "a1f0je01O000000000000000000001O000001O00000000000[OdZO6\\e0Je000000000000001O00000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 323_kTQ8zZdbuW0_00000217 | 323_kTQ8zZdbuW0_00000217.jpg | {
"data_source": "VIPSeg",
"file_name": "323_kTQ8zZdbuW0_00000217.jpg",
"height": 720,
"id": "323_kTQ8zZdbuW0_00000217",
"width": 1280
} | {
"caption": "A person in yellow shorts is captured mid-air, diving headfirst into the sea from a tall, dark rocky cliff, with other rock formations visible in the background and a hazy sky above.",
"caption_ann": "A <3:person in yellow shorts> is captured mid-air, diving headfirst into the <2:sea> from a <1:tall, dark rocky cliff>, with other <1:rock formations> visible in the background and a <0:hazy sky> above.",
"id": 1791,
"image_id": "323_kTQ8zZdbuW0_00000217",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "person in yellow shorts"
},
{
"mask_ids": [
2
],
"txt_desc": "sea"
},
{
"mask_ids": [
1
],
"txt_desc": "tall, dark rocky cliff"
},
{
"mask_ids": [
1
],
"txt_desc": "rock formations"
},
{
"mask_ids": [
0
],
"txt_desc": "hazy sky"
}
],
"labels": [
"sky",
"mountain",
"sea",
"person"
]
} | [
{
"area": 14880,
"bbox": [
998,
0,
282,
112
],
"category_id": 28,
"id": 19405,
"image_id": "323_kTQ8zZdbuW0_00000217",
"iscrowd": 0,
"segmentation": {
"counts": "Pgme01_f0001O00001O00001O0000001O0000001O00001O00001O001O001O001O001O1O00001O00001O0000001O0... | [
{
"id": 1,
"name": "object"
}
] |
train | 327_N8v-fBj-x2s_00000317 | 327_N8v-fBj-x2s_00000317.jpg | {
"data_source": "VIPSeg",
"file_name": "327_N8v-fBj-x2s_00000317.jpg",
"height": 720,
"id": "327_N8v-fBj-x2s_00000317",
"width": 1280
} | {
"caption": "A bright orange and gray tent is pitched on the sandy ground of a vast desert landscape, with its door open. Next to the tent, a touring bicycle is parked, suggesting a bike-packing adventure. Two dark-colored bags or panniers are placed on the sand behind the tent. The campsite is surrounded by sparse desert scrub and bushes and distant green plants and bushes creates a horizon. In the background, a large mountain range is silhouetted against the fading light of the evening sky.",
"caption_ann": "A <5:bright orange and gray tent> is pitched on the <0:sandy ground of a vast desert landscape>, with its door open. Next to the <5:tent>, a <6:touring bicycle> is parked, suggesting a bike-packing adventure. <7,8:Two dark-colored bags or panniers> are placed on the sand behind the <6:tent>. The campsite is surrounded by sparse <4:desert scrub and bushes> and <1:distant green plants and bushes> creates a horizon. In the background, a <3:large mountain range> is silhouetted against the <2:fading light of the evening sky>.",
"id": 1792,
"image_id": "327_N8v-fBj-x2s_00000317",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "bright orange and gray tent"
},
{
"mask_ids": [
0
],
"txt_desc": "sandy ground of a vast desert landscape"
},
{
"mask_ids": [
5
],
"txt_desc": "tent"
},
{
"mask_ids": [
6
],
"txt_desc": "touring bicycle"
},
{
"mask_ids": [
7,
8
],
"txt_desc": "Two dark-colored bags or panniers"
},
{
"mask_ids": [
6
],
"txt_desc": "tent"
},
{
"mask_ids": [
4
],
"txt_desc": "desert scrub and bushes"
},
{
"mask_ids": [
1
],
"txt_desc": "distant green plants and bushes"
},
{
"mask_ids": [
3
],
"txt_desc": "large mountain range"
},
{
"mask_ids": [
2
],
"txt_desc": "fading light of the evening sky"
}
],
"labels": [
"ground",
"grass",
"sky",
"mountain",
"other_plant",
"tent",
"bicycle",
"bag_or_package",
"bag_or_package"
]
} | [
{
"area": 368852,
"bbox": [
0,
191,
1280,
529
],
"category_id": 14,
"id": 19409,
"image_id": "327_N8v-fBj-x2s_00000317",
"iscrowd": 0,
"segmentation": {
"counts": "Y=W9Z=O001O00001O1O00001O1O00001O1O0000001O1O00001O1O00001O001O001O001O001O001O00001O1O000... | [
{
"id": 1,
"name": "object"
}
] |
train | 329_2vubTJvigGA_00000205 | 329_2vubTJvigGA_00000205.jpg | {
"data_source": "VIPSeg",
"file_name": "329_2vubTJvigGA_00000205.jpg",
"height": 720,
"id": "329_2vubTJvigGA_00000205",
"width": 1280
} | {
"caption": "A woman with long hair stands in front of a white closet with open doors that is filled with clothes, various boxes, a traveling case, and several baskets, while holding white flowers. The space features white walls, a window with daylight streaming through, a white door frame opening to another room with another door on the left.",
"caption_ann": "A <9:woman with long hair> stands in front of a <2:white closet> with <6,7: open doors> that is filled with <3:clothes>, <10,11,12,13,14:various boxes>, a <15:traveling case>, and <16,17,18,19:several baskets>, while holding <1:white flowers>. The space features <0:white walls>, a <8:window> with daylight streaming through, a <5:white door frame> opening to another room with another <4:door> on the left.",
"id": 1793,
"image_id": "329_2vubTJvigGA_00000205",
"label_matched": [
{
"mask_ids": [
9
],
"txt_desc": "woman with long hair"
},
{
"mask_ids": [
2
],
"txt_desc": "white closet"
},
{
"mask_ids": [
6,
7
],
"txt_desc": "open doors"
},
{
"mask_ids": [
3
],
"txt_desc": "clothes"
},
{
"mask_ids": [
10,
11,
12,
13,
14
],
"txt_desc": "various boxes"
},
{
"mask_ids": [
15
],
"txt_desc": "traveling case"
},
{
"mask_ids": [
16,
17,
18,
19
],
"txt_desc": "several baskets"
},
{
"mask_ids": [
1
],
"txt_desc": "white flowers"
},
{
"mask_ids": [
0
],
"txt_desc": "white walls"
},
{
"mask_ids": [
8
],
"txt_desc": "window"
},
{
"mask_ids": [
5
],
"txt_desc": "white door frame"
},
{
"mask_ids": [
4
],
"txt_desc": "door"
}
],
"labels": [
"wall",
"flower",
"cupboard_or_showcase_or_storage_rack",
"clothes",
"door",
"door",
"door",
"door",
"window",
"person",
"box",
"box",
"box",
"box",
"box",
"traveling_case_or_trolley_case",
"basket",
"basket",
"basket",
"basket"
]
} | [
{
"area": 378995,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 19418,
"image_id": "329_2vubTJvigGA_00000205",
"iscrowd": 0,
"segmentation": {
"counts": "0\\1S1V2m=ZORBf0n=ZORBf0n=ZORBf0n=ZORBf0n=ZORBf0n=ZORBf0n=ZOSBe0m=[OSBe0m=[OSBe0m=[OSBe0m=[OS... | [
{
"id": 1,
"name": "object"
}
] |
train | 32_sLFZe-epD3M_00000241 | 32_sLFZe-epD3M_00000241.jpg | {
"data_source": "VIPSeg",
"file_name": "32_sLFZe-epD3M_00000241.jpg",
"height": 720,
"id": "32_sLFZe-epD3M_00000241",
"width": 1280
} | {
"caption": "In a cozy bedroom, a fawn-colored bulldog sits with its back to the camera on a bed with white coverlet. The dog is focused on a large television screen that is displaying a vibrant, natural scene. The television rests on a dark wood media console. The room has dark-colored walls and a slanted white ceiling. To the right of the television, a framed picture containing three smaller images hangs on the wall.",
"caption_ann": "In a cozy bedroom, a <3:fawn-colored bulldog> sits with its back to the camera on a <4:bed with white coverlet>. The <3:dog> is focused on a <6:large television screen> that is displaying a vibrant, natural scene. The <6:television> rests on a <2:dark wood media console>. The room has <0:dark-colored walls> and a <1:slanted white ceiling>. To the right of the television, a <5:framed picture containing three smaller images> hangs on the <0:wall>.",
"id": 1794,
"image_id": "32_sLFZe-epD3M_00000241",
"label_matched": [
{
"mask_ids": [
3
],
"txt_desc": "fawn-colored bulldog"
},
{
"mask_ids": [
4
],
"txt_desc": "bed with white coverlet"
},
{
"mask_ids": [
3
],
"txt_desc": "dog"
},
{
"mask_ids": [
6
],
"txt_desc": "large television screen"
},
{
"mask_ids": [
6
],
"txt_desc": "television"
},
{
"mask_ids": [
2
],
"txt_desc": "dark wood media console"
},
{
"mask_ids": [
0
],
"txt_desc": "dark-colored walls"
},
{
"mask_ids": [
1
],
"txt_desc": "slanted white ceiling"
},
{
"mask_ids": [
5
],
"txt_desc": "framed picture containing three smaller images"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
}
],
"labels": [
"wall",
"ceiling",
"cupboard_or_showcase_or_storage_rack",
"dog",
"bed",
"painting_or_poster",
"screen_or_television"
]
} | [
{
"area": 253910,
"bbox": [
0,
0,
1280,
672
],
"category_id": 0,
"id": 19438,
"image_id": "32_sLFZe-epD3M_00000241",
"iscrowd": 0,
"segmentation": {
"counts": "^3ba0n4O100000000000001O000O1000000000001O00000O10000000001O0000000O100000001O000000000O100000... | [
{
"id": 1,
"name": "object"
}
] |
train | 330_QuvULB8itOg_00000917 | 330_QuvULB8itOg_00000917.jpg | {
"data_source": "VIPSeg",
"file_name": "330_QuvULB8itOg_00000917.jpg",
"height": 720,
"id": "330_QuvULB8itOg_00000917",
"width": 1280
} | {
"caption": "A man in a dark suit,light pink shirt and a patterned tie stands on a floor designed with stone patio, gesturing towards a set of two bronze sculptures. The sculptures depict a woman in short hair seated on bench and a nun, seated on bench engaged in conversation. The three stone benches arranged in a circular pattern on a stone patio. An empty stone bench is also part of the arrangement on the patio. The entire scene is set in a landscaped courtyard with green grass, leafy trees and a paved walkway crossing horizontally on the field of grass . In the background, a large, multi-story brick building is visible, along with a lamppost and a small partially visible object. The scene is set under a bright, overcast sky.",
"caption_ann": "A <10:man in a dark suit,light pink shirt and a patterned tie> stands on a <1: floor designed with stone patio>, gesturing towards a set of <8,9:two bronze sculptures>. The sculptures depict <8:a woman in short hair> seated on <11:bench> and <9:a nun>, seated on <13:bench> engaged in conversation. The <11,12,13:three stone benches> arranged in a circular pattern on a stone patio. An empty <12:stone bench> is also part of the arrangement on the patio. The entire scene is set in a landscaped courtyard with <2:green grass>, <7:leafy trees> and a <3:paved walkway> crossing horizontally on the <2:field of grass> . In the background, a large, <4:multi-story brick building> is visible, along with a <0:lamppost> and a <6:small partially visible object>. The scene is set under a <5:bright, overcast sky>.",
"id": 1795,
"image_id": "330_QuvULB8itOg_00000917",
"label_matched": [
{
"mask_ids": [
10
],
"txt_desc": "man in a dark suit,light pink shirt and a patterned tie"
},
{
"mask_ids": [
1
],
"txt_desc": "floor designed with stone patio"
},
{
"mask_ids": [
8,
9
],
"txt_desc": "two bronze sculptures"
},
{
"mask_ids": [
8
],
"txt_desc": "a woman in short hair"
},
{
"mask_ids": [
11
],
"txt_desc": "bench"
},
{
"mask_ids": [
9
],
"txt_desc": "a nun"
},
{
"mask_ids": [
13
],
"txt_desc": "bench"
},
{
"mask_ids": [
11,
12,
13
],
"txt_desc": "three stone benches"
},
{
"mask_ids": [
12
],
"txt_desc": "stone bench"
},
{
"mask_ids": [
2
],
"txt_desc": "green grass"
},
{
"mask_ids": [
7
],
"txt_desc": "leafy trees"
},
{
"mask_ids": [
3
],
"txt_desc": "paved walkway"
},
{
"mask_ids": [
2
],
"txt_desc": "field of grass"
},
{
"mask_ids": [
4
],
"txt_desc": "multi-story brick building"
},
{
"mask_ids": [
0
],
"txt_desc": "lamppost"
},
{
"mask_ids": [
6
],
"txt_desc": "small partially visible object"
},
{
"mask_ids": [
5
],
"txt_desc": "bright, overcast sky"
}
],
"labels": [
"pole",
"ground",
"grass",
"path",
"building",
"sky",
"billboard_or_Bulletin_Board",
"tree",
"sculpture",
"sculpture",
"person",
"bench",
"bench",
"bench"
]
} | [
{
"area": 1260,
"bbox": [
289,
295,
18,
70
],
"category_id": 12,
"id": 19445,
"image_id": "330_QuvULB8itOg_00000917",
"iscrowd": 0,
"segmentation": {
"counts": "g_[6V2Zd0000000000000000000000000000000000Yk[e0",
"size": [
720,
1280
... | [
{
"id": 1,
"name": "object"
}
] |
train | 335_HW5qLhdqk-g_00004435 | 335_HW5qLhdqk-g_00004435.jpg | {
"data_source": "VIPSeg",
"file_name": "335_HW5qLhdqk-g_00004435.jpg",
"height": 720,
"id": "335_HW5qLhdqk-g_00004435",
"width": 1280
} | {
"caption": "A young woman with blonde hair, wearing a black hoodie, poses for a selfie. She is holding up a small black digital camera to capture her reflection on the mirror. The mirror's reflection shows her room, which has grey walls and a white ceiling with a light fixture. A window is visible, framed by dark grey curtains. To the right is a well-organized open closet system with drawers and hanging clothes. A Dark gray boxes sit on top of the closet unit. A large rounded mirror is mounted on the <0:wall towards the left side.",
"caption_ann": "A <8:young woman with blonde hair, wearing a black hoodie,> poses for a selfie. She is holding up a <6:small black digital camera> to capture her reflection on the mirror. The mirror's reflection shows her room, which has <0:grey walls> and a <1:white ceiling> with a <2:light fixture>. A <7:window> is visible, framed by <4:dark grey curtains>. To the right is a <3:well-organized open closet system> with drawers and <5:hanging clothes>. A <9:Dark gray boxes> sit on top of the closet unit. A <10:large rounded mirror is mounted on the <0:wall> towards the left side.",
"id": 1796,
"image_id": "335_HW5qLhdqk-g_00004435",
"label_matched": [
{
"mask_ids": [
8
],
"txt_desc": "young woman with blonde hair, wearing a black hoodie,"
},
{
"mask_ids": [
6
],
"txt_desc": "small black digital camera"
},
{
"mask_ids": [
0
],
"txt_desc": "grey walls"
},
{
"mask_ids": [
1
],
"txt_desc": "white ceiling"
},
{
"mask_ids": [
2
],
"txt_desc": "light fixture"
},
{
"mask_ids": [
7
],
"txt_desc": "window"
},
{
"mask_ids": [
4
],
"txt_desc": "dark grey curtains"
},
{
"mask_ids": [
3
],
"txt_desc": "well-organized open closet system"
},
{
"mask_ids": [
5
],
"txt_desc": "hanging clothes"
},
{
"mask_ids": [
9
],
"txt_desc": "Dark gray boxes"
},
{
"mask_ids": [
10
],
"txt_desc": "large rounded mirror is mounted on the <0:wall"
}
],
"labels": [
"wall",
"ceiling",
"lamp",
"cupboard_or_showcase_or_storage_rack",
"curtain",
"clothes",
"other_electronic_product",
"window",
"person",
"box",
"mirror"
]
} | [
{
"area": 293681,
"bbox": [
0,
0,
773,
720
],
"category_id": 0,
"id": 19459,
"image_id": "335_HW5qLhdqk-g_00004435",
"iscrowd": 0,
"segmentation": {
"counts": "0kf`76dg_H_2gFYNa8g1_GdNV8^1gG[Oa7e0_HFV7:iH?_6B`Ij0T6WOkI`1^5`NbJl1R5UNmJb2\\4^MdKn2[OaKROb1a... | [
{
"id": 1,
"name": "object"
}
] |
train | 339_pD7kE8qUz5A_00005252 | 339_pD7kE8qUz5A_00005252.jpg | {
"data_source": "VIPSeg",
"file_name": "339_pD7kE8qUz5A_00005252.jpg",
"height": 720,
"id": "339_pD7kE8qUz5A_00005252",
"width": 1280
} | {
"caption": "During a christening ceremony, a priest, dressed in white and gold vestments, stands and reads from a book with a black cover. In the pews, a woman in a black dress holds a baby dressed in a traditional white christening gown and bonnet. Seated beside them towards the left is a attentive man in a dark gray suit and tie, and towards the right is a young woman with blonde hair. Another man wearing a gray suit, white shirt, and diagonally striped tie sits on the extreme right. The family is seated on a wooden church pew, with two other pews, also visible. The church has a carpeted floor and light-colored brick walls. On the wall hangs a partially visible religious painting. In the background, there are two wooden doors with gold trim located on both sides.",
"caption_ann": "During a christening ceremony, a <5:priest, dressed in white and gold vestments,> stands and reads from a <2:book with a black cover>. In the pews, a <7:woman in a black dress> holds a <6:baby dressed in a traditional white christening gown and bonnet>. Seated beside them towards the left is a <9:attentive man in a dark gray suit and tie>, and towards the right is a <8:young woman with blonde hair>. Another <10:man wearing a gray suit, white shirt, and diagonally striped tie> sits on the extreme right. The family is seated on a <13:wooden church pew>, with <11,12:two other pews>, also visible. The church has a <1:carpeted floor> and <0:light-colored brick walls>. On the <0:wall> hangs a <14:partially visible religious painting>. In the background, there are <3,4:two wooden doors with gold trim> located on both sides.",
"id": 1797,
"image_id": "339_pD7kE8qUz5A_00005252",
"label_matched": [
{
"mask_ids": [
5
],
"txt_desc": "priest, dressed in white and gold vestments,"
},
{
"mask_ids": [
2
],
"txt_desc": "book with a black cover"
},
{
"mask_ids": [
7
],
"txt_desc": "woman in a black dress"
},
{
"mask_ids": [
6
],
"txt_desc": "baby dressed in a traditional white christening gown and bonnet"
},
{
"mask_ids": [
9
],
"txt_desc": "attentive man in a dark gray suit and tie"
},
{
"mask_ids": [
8
],
"txt_desc": "young woman with blonde hair"
},
{
"mask_ids": [
10
],
"txt_desc": "man wearing a gray suit, white shirt, and diagonally striped tie"
},
{
"mask_ids": [
13
],
"txt_desc": "wooden church pew"
},
{
"mask_ids": [
11,
12
],
"txt_desc": "two other pews"
},
{
"mask_ids": [
1
],
"txt_desc": "carpeted floor"
},
{
"mask_ids": [
0
],
"txt_desc": "light-colored brick walls"
},
{
"mask_ids": [
0
],
"txt_desc": "wall"
},
{
"mask_ids": [
14
],
"txt_desc": "partially visible religious painting"
},
{
"mask_ids": [
3,
4
],
"txt_desc": "two wooden doors with gold trim"
}
],
"labels": [
"wall",
"floor",
"book",
"door",
"door",
"person",
"person",
"person",
"person",
"person",
"person",
"bench",
"bench",
"bench",
"painting_or_poster"
]
} | [
{
"area": 242231,
"bbox": [
0,
0,
1068,
509
],
"category_id": 0,
"id": 19470,
"image_id": "339_pD7kE8qUz5A_00005252",
"iscrowd": 0,
"segmentation": {
"counts": "0m?c600000000000000000000000000000000000000000000000000000000O1000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 33_qEvYDmF1Lx4_00000190 | 33_qEvYDmF1Lx4_00000190.jpg | {
"data_source": "VIPSeg",
"file_name": "33_qEvYDmF1Lx4_00000190.jpg",
"height": 720,
"id": "33_qEvYDmF1Lx4_00000190",
"width": 1280
} | {
"caption": "A yellow and orange ambulance is parked on a vast snowfield under a pale sky, where a group of seven people in heavy winter gear are gathered. Among the group, a child in a blue jacket is bent over looking at the ice, near a child in a pink snowsuit. Meanwhile, two people wearing a bright blue and dark blue jacket walk across the ice, and a person in a dark jacket looks towards the ambulance.",
"caption_ann": "A <2:yellow and orange ambulance> is parked on a vast <1:snowfield> under a <0:pale sky>, where a group of <3,4,5,6,7,8,9:seven people in heavy winter gear> are gathered. Among the group, a <5:child in a blue jacket> is bent over looking at the ice, near a <9:child in a pink snowsuit>. Meanwhile, <7,8:two people wearing a bright blue and dark blue jacket> walk across the ice, and a <6:person in a dark jacket> looks towards the ambulance.",
"id": 1798,
"image_id": "33_qEvYDmF1Lx4_00000190",
"label_matched": [
{
"mask_ids": [
2
],
"txt_desc": "yellow and orange ambulance"
},
{
"mask_ids": [
1
],
"txt_desc": "snowfield"
},
{
"mask_ids": [
0
],
"txt_desc": "pale sky"
},
{
"mask_ids": [
3,
4,
5,
6,
7,
8,
9
],
"txt_desc": "seven people in heavy winter gear"
},
{
"mask_ids": [
5
],
"txt_desc": "child in a blue jacket"
},
{
"mask_ids": [
9
],
"txt_desc": "child in a pink snowsuit"
},
{
"mask_ids": [
7,
8
],
"txt_desc": "two people wearing a bright blue and dark blue jacket"
},
{
"mask_ids": [
6
],
"txt_desc": "person in a dark jacket"
}
],
"labels": [
"sky",
"snowfield",
"car",
"person",
"person",
"person",
"person",
"person",
"person",
"person"
]
} | [
{
"area": 120180,
"bbox": [
0,
0,
1280,
113
],
"category_id": 28,
"id": 19485,
"image_id": "33_qEvYDmF1Lx4_00000190",
"iscrowd": 0,
"segmentation": {
"counts": "0a3ob0O10000000000000000000000000000000000000000000000000000000000000000000000000000000000000... | [
{
"id": 1,
"name": "object"
}
] |
train | 343_2c-SvLO08ZI_00004367 | 343_2c-SvLO08ZI_00004367.jpg | {
"data_source": "VIPSeg",
"file_name": "343_2c-SvLO08ZI_00004367.jpg",
"height": 720,
"id": "343_2c-SvLO08ZI_00004367",
"width": 1280
} | {
"caption": "The image displays a medical clinical setting. two persons are dressed in surgical scrubs, caps, and masks. The woman on the right, wearing a brown scrub top and glasses, is gesturing with her hands as if explaining a procedure. The person on the left wearing blue disposable surgical gown is seen from behind. A medical bed covered with a blue sheet, is visible on the left side. The room has plain, light-colored walls and a dark colored floor.",
"caption_ann": "The image displays a medical clinical setting. <2,3:two persons are dressed in surgical scrubs, caps, and masks>. The <3:woman on the right, wearing a brown scrub top and glasses>, is gesturing with her hands as if explaining a procedure. The <2:person on the left wearing blue disposable surgical gown> is seen from behind. A <4:medical bed covered with a blue sheet,> is visible on the left side. The room has <0:plain, light-colored walls> and a <1:dark colored floor>.",
"id": 1799,
"image_id": "343_2c-SvLO08ZI_00004367",
"label_matched": [
{
"mask_ids": [
2,
3
],
"txt_desc": "two persons are dressed in surgical scrubs, caps, and masks"
},
{
"mask_ids": [
3
],
"txt_desc": "woman on the right, wearing a brown scrub top and glasses"
},
{
"mask_ids": [
2
],
"txt_desc": "person on the left wearing blue disposable surgical gown"
},
{
"mask_ids": [
4
],
"txt_desc": "medical bed covered with a blue sheet,"
},
{
"mask_ids": [
0
],
"txt_desc": "plain, light-colored walls"
},
{
"mask_ids": [
1
],
"txt_desc": "dark colored floor"
}
],
"labels": [
"wall",
"floor",
"person",
"person",
"bed"
]
} | [
{
"area": 403910,
"bbox": [
0,
0,
1280,
720
],
"category_id": 0,
"id": 19495,
"image_id": "343_2c-SvLO08ZI_00004367",
"iscrowd": 0,
"segmentation": {
"counts": "0S8]>000000000000000000000000000000000000000000000000000000000000000000000000N200M3O1N2O1N2O1... | [
{
"id": 1,
"name": "object"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.