Dataset Preview
Viewer
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
An error occurred while generating the dataset All the data files must have the same columns, but at some point there are 4 missing columns ({'label', 'bbox', 'color', 'occlusion_percentage'}) This happened while the json dataset builder was generating data using <datasets.download.streaming_download_manager.FilesIterable object at 0x7f95e01bce50> Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)
Error code:   UnexpectedError

Need help to make the dataset viewer work? Open a discussion for direct support.

label
string
color
sequence
occlusion_percentage
float64
bbox
dict
car
[ 142, 48, 13, 255 ]
0.04
{ "xMax": 276, "xMin": 0, "yMax": 684, "yMin": 373 }
person
[ 114, 47, 11, 255 ]
0.62
{ "xMax": 1095, "xMin": 997, "yMax": 703, "yMin": 439 }
person
[ 208, 151, 11, 255 ]
28.36
{ "xMax": 310, "xMin": 285, "yMax": 535, "yMin": 455 }
car
[ 211, 40, 13, 255 ]
62.02
{ "xMax": 412, "xMin": 335, "yMax": 521, "yMin": 475 }
person
[ 230, 218, 11, 255 ]
94.82
{ "xMax": 561, "xMin": 524, "yMax": 558, "yMin": 470 }
person
[ 231, 63, 11, 255 ]
91.05
{ "xMax": 1641, "xMin": 1620, "yMax": 470, "yMin": 416 }
person
[ 119, 136, 11, 255 ]
7.68
{ "xMax": 1403, "xMin": 1380, "yMax": 496, "yMin": 428 }
person
[ 187, 28, 11, 255 ]
54.14
{ "xMax": 203, "xMin": 81, "yMax": 747, "yMin": 465 }
car
[ 235, 52, 13, 255 ]
98.47
{ "xMax": 1235, "xMin": 1142, "yMax": 515, "yMin": 453 }
person
[ 210, 195, 11, 255 ]
14.37
{ "xMax": 1460, "xMin": 1411, "yMax": 556, "yMin": 395 }
motorcycle
[ 25, 32, 17, 255 ]
51.17
{ "xMax": 1267, "xMin": 1078, "yMax": 640, "yMin": 482 }
car
[ 212, 140, 13, 255 ]
99.27
{ "xMax": 1955, "xMin": 1760, "yMax": 480, "yMin": 389 }
motorcycle
[ 2, 120, 17, 255 ]
43.66
{ "xMax": 1092, "xMin": 841, "yMax": 684, "yMin": 497 }
motorcycle
[ 164, 116, 17, 255 ]
0
{ "xMax": 971, "xMin": 680, "yMax": 719, "yMin": 519 }
motorcycle
[ 96, 224, 17, 255 ]
58.6
{ "xMax": 1290, "xMin": 1255, "yMax": 519, "yMin": 464 }
rider
[ 96, 224, 12, 255 ]
47.1
{ "xMax": 1289, "xMin": 1253, "yMax": 505, "yMin": 440 }
person
[ 24, 187, 11, 255 ]
7.99
{ "xMax": 1249, "xMin": 1236, "yMax": 493, "yMin": 451 }
person
[ 140, 4, 11, 255 ]
0
{ "xMax": 1759, "xMin": 1704, "yMax": 513, "yMin": 375 }
car
[ 3, 220, 13, 255 ]
66.86
{ "xMax": 1388, "xMin": 1332, "yMax": 487, "yMin": 450 }
person
[ 233, 8, 11, 255 ]
7.86
{ "xMax": 1717, "xMin": 1677, "yMax": 512, "yMin": 386 }
person
[ 93, 179, 11, 255 ]
59.26
{ "xMax": 1553, "xMin": 1539, "yMax": 474, "yMin": 427 }
motorcycle
[ 141, 203, 17, 255 ]
44.02
{ "xMax": 1198, "xMin": 959, "yMax": 659, "yMin": 482 }
car
[ 142, 48, 13, 255 ]
81.05
{ "xMax": 1581, "xMin": 1450, "yMax": 531, "yMin": 465 }
car
[ 95, 124, 13, 255 ]
0.53
{ "xMax": 901, "xMin": 784, "yMax": 634, "yMin": 532 }
person
[ 45, 55, 11, 255 ]
0.02
{ "xMax": 1778, "xMin": 1665, "yMax": 667, "yMin": 346 }
motorcycle
[ 71, 112, 17, 255 ]
69.75
{ "xMax": 1463, "xMin": 1432, "yMax": 517, "yMin": 492 }
person
[ 1, 175, 11, 255 ]
46.49
{ "xMax": 1136, "xMin": 1124, "yMax": 546, "yMin": 517 }
car
[ 50, 144, 13, 255 ]
81.35
{ "xMax": 1414, "xMin": 1281, "yMax": 534, "yMin": 485 }
person
[ 26, 132, 11, 255 ]
57.71
{ "xMax": 1350, "xMin": 1336, "yMax": 534, "yMin": 489 }
person
[ 43, 110, 11, 255 ]
45.83
{ "xMax": 698, "xMin": 670, "yMax": 611, "yMin": 565 }
person
[ 182, 194, 11, 255 ]
76.72
{ "xMax": 761, "xMin": 738, "yMax": 609, "yMin": 562 }
person
[ 162, 71, 11, 255 ]
50.5
{ "xMax": 1282, "xMin": 1272, "yMax": 534, "yMin": 500 }
person
[ 252, 30, 11, 255 ]
86.49
{ "xMax": 870, "xMin": 842, "yMax": 617, "yMin": 535 }
motorcycle
[ 2, 120, 17, 255 ]
83.47
{ "xMax": 675, "xMin": 617, "yMax": 633, "yMin": 589 }
motorcycle
[ 164, 116, 17, 255 ]
95.73
{ "xMax": 705, "xMin": 644, "yMax": 631, "yMin": 587 }
car
[ 165, 216, 13, 255 ]
29.28
{ "xMax": 1207, "xMin": 1171, "yMax": 543, "yMin": 512 }
person
[ 2, 120, 11, 255 ]
60.54
{ "xMax": 1349, "xMin": 1313, "yMax": 578, "yMin": 466 }
train
[ 210, 195, 16, 255 ]
0.12
{ "xMax": 472, "xMin": 0, "yMax": 969, "yMin": 212 }
person
[ 229, 118, 11, 255 ]
28.08
{ "xMax": 1400, "xMin": 1360, "yMax": 586, "yMin": 458 }
person
[ 159, 26, 11, 255 ]
0.56
{ "xMax": 945, "xMin": 903, "yMax": 605, "yMin": 525 }
person
[ 73, 56, 11, 255 ]
77.45
{ "xMax": 1020, "xMin": 1007, "yMax": 565, "yMin": 533 }
person
[ 71, 112, 11, 255 ]
51.24
{ "xMax": 656, "xMin": 636, "yMax": 632, "yMin": 564 }
person
[ 49, 44, 11, 255 ]
0
{ "xMax": 1470, "xMin": 1437, "yMax": 565, "yMin": 456 }
person
[ 20, 198, 11, 255 ]
0.09
{ "xMax": 1170, "xMin": 1121, "yMax": 577, "yMin": 497 }
car
[ 118, 36, 13, 255 ]
9.39
{ "xMax": 1162, "xMin": 1029, "yMax": 582, "yMin": 522 }
car
[ 189, 228, 13, 255 ]
83.85
{ "xMax": 966, "xMin": 838, "yMax": 588, "yMin": 540 }
person
[ 70, 167, 11, 255 ]
48.57
{ "xMax": 1036, "xMin": 1028, "yMax": 559, "yMin": 532 }
person
[ 211, 40, 11, 255 ]
6.81
{ "xMax": 1087, "xMin": 1048, "yMax": 548, "yMin": 445 }
person
[ 118, 36, 11, 255 ]
0.26
{ "xMax": 604, "xMin": 523, "yMax": 583, "yMin": 355 }
person
[ 239, 75, 11, 255 ]
0.46
{ "xMax": 2047, "xMin": 2024, "yMax": 675, "yMin": 574 }
person
[ 141, 203, 11, 255 ]
76.19
{ "xMax": 1071, "xMin": 1041, "yMax": 545, "yMin": 451 }
person
[ 253, 130, 11, 255 ]
2.57
{ "xMax": 1141, "xMin": 1117, "yMax": 540, "yMin": 473 }
person
[ 137, 214, 11, 255 ]
61.57
{ "xMax": 1159, "xMin": 1136, "yMax": 532, "yMin": 480 }
car
[ 142, 48, 13, 255 ]
85.19
{ "xMax": 1552, "xMin": 1482, "yMax": 564, "yMin": 531 }
train
[ 16, 15, 16, 255 ]
60.87
{ "xMax": 1383, "xMin": 1270, "yMax": 569, "yMin": 438 }
person
[ 229, 118, 11, 255 ]
0.31
{ "xMax": 1481, "xMin": 1470, "yMax": 564, "yMin": 522 }
person
[ 234, 208, 11, 255 ]
5.58
{ "xMax": 162, "xMin": 0, "yMax": 868, "yMin": 55 }
person
[ 226, 119, 11, 255 ]
90.81
{ "xMax": 1751, "xMin": 1709, "yMax": 620, "yMin": 534 }
person
[ 95, 124, 11, 255 ]
6.51
{ "xMax": 1614, "xMin": 1596, "yMax": 589, "yMin": 531 }
person
[ 164, 116, 11, 255 ]
88.64
{ "xMax": 1123, "xMin": 1107, "yMax": 521, "yMin": 475 }
person
[ 49, 44, 11, 255 ]
36.02
{ "xMax": 1503, "xMin": 1492, "yMax": 565, "yMin": 525 }
car
[ 234, 208, 13, 255 ]
0.61
{ "xMax": 1599, "xMin": 1489, "yMax": 602, "yMin": 528 }
person
[ 138, 59, 11, 255 ]
43.05
{ "xMax": 1414, "xMin": 1403, "yMax": 546, "yMin": 517 }
car
[ 73, 56, 13, 255 ]
0.31
{ "xMax": 1387, "xMin": 1132, "yMax": 652, "yMin": 451 }
car
[ 55, 71, 13, 255 ]
0.27
{ "xMax": 1936, "xMin": 1669, "yMax": 681, "yMin": 534 }
person
[ 2, 120, 11, 255 ]
18.48
{ "xMax": 1458, "xMin": 1442, "yMax": 559, "yMin": 517 }
person
[ 26, 132, 11, 255 ]
49.61
{ "xMax": 1421, "xMin": 1414, "yMax": 547, "yMin": 520 }
person
[ 137, 214, 11, 255 ]
39.22
{ "xMax": 1802, "xMin": 1739, "yMax": 630, "yMin": 477 }
person
[ 72, 212, 11, 255 ]
56.82
{ "xMax": 1238, "xMin": 1229, "yMax": 560, "yMin": 533 }
train
[ 210, 195, 16, 255 ]
0.55
{ "xMax": 1437, "xMin": 1223, "yMax": 634, "yMin": 403 }
person
[ 141, 203, 11, 255 ]
21.3
{ "xMax": 1644, "xMin": 1624, "yMax": 577, "yMin": 500 }
person
[ 102, 51, 11, 255 ]
73.18
{ "xMax": 1177, "xMin": 1163, "yMax": 573, "yMin": 527 }
person
[ 234, 208, 11, 255 ]
0
{ "xMax": 1007, "xMin": 978, "yMax": 603, "yMin": 522 }
person
[ 222, 102, 11, 255 ]
64.16
{ "xMax": 1548, "xMin": 1532, "yMax": 558, "yMin": 513 }
car
[ 234, 208, 13, 255 ]
0.52
{ "xMax": 1125, "xMin": 1014, "yMax": 622, "yMin": 527 }
car
[ 222, 102, 13, 255 ]
99.59
{ "xMax": 1490, "xMin": 1439, "yMax": 562, "yMin": 525 }
person
[ 211, 40, 11, 255 ]
26.9
{ "xMax": 1630, "xMin": 1601, "yMax": 575, "yMin": 496 }
person
[ 253, 23, 11, 255 ]
38.99
{ "xMax": 1182, "xMin": 1164, "yMax": 573, "yMin": 524 }
person
[ 45, 55, 11, 255 ]
0
{ "xMax": 711, "xMin": 641, "yMax": 658, "yMin": 516 }
person
[ 158, 100, 11, 255 ]
38.46
{ "xMax": 1519, "xMin": 1506, "yMax": 552, "yMin": 518 }
person
[ 95, 124, 11, 255 ]
0
{ "xMax": 779, "xMin": 742, "yMax": 639, "yMin": 525 }
person
[ 28, 238, 11, 255 ]
85.71
{ "xMax": 1533, "xMin": 1525, "yMax": 553, "yMin": 518 }
car
[ 73, 56, 13, 255 ]
0.58
{ "xMax": 1239, "xMin": 1167, "yMax": 588, "yMin": 526 }
car
[ 212, 183, 13, 255 ]
1.69
{ "xMax": 1510, "xMin": 1433, "yMax": 572, "yMin": 518 }
person
[ 49, 44, 11, 255 ]
57.58
{ "xMax": 1520, "xMin": 1513, "yMax": 549, "yMin": 521 }
person
[ 253, 130, 11, 255 ]
0
{ "xMax": 1673, "xMin": 1636, "yMax": 596, "yMin": 497 }
person
[ 155, 96, 11, 255 ]
90.2
{ "xMax": 1231, "xMin": 1218, "yMax": 567, "yMin": 530 }
person
[ 113, 202, 11, 255 ]
0.08
{ "xMax": 2015, "xMin": 1940, "yMax": 689, "yMin": 484 }
person
[ 118, 36, 11, 255 ]
68.46
{ "xMax": 1002, "xMin": 982, "yMax": 595, "yMin": 533 }
person
[ 25, 143, 11, 255 ]
90.09
{ "xMax": 1632, "xMin": 1618, "yMax": 561, "yMin": 512 }
person
[ 1, 175, 11, 255 ]
86.39
{ "xMax": 1492, "xMin": 1476, "yMax": 520, "yMin": 475 }
motorcycle
[ 164, 116, 17, 255 ]
69.38
{ "xMax": 615, "xMin": 389, "yMax": 604, "yMin": 470 }
person
[ 209, 95, 11, 255 ]
83.2
{ "xMax": 1423, "xMin": 1409, "yMax": 516, "yMin": 477 }
person
[ 165, 216, 11, 255 ]
24.2
{ "xMax": 697, "xMin": 674, "yMax": 512, "yMin": 438 }
person
[ 136, 114, 11, 255 ]
30.07
{ "xMax": 406, "xMin": 360, "yMax": 510, "yMin": 419 }
motorcycle
[ 141, 203, 17, 255 ]
75.3
{ "xMax": 503, "xMin": 301, "yMax": 581, "yMin": 453 }
person
[ 71, 112, 11, 255 ]
10.45
{ "xMax": 228, "xMin": 125, "yMax": 652, "yMin": 341 }
person
[ 255, 75, 11, 255 ]
29.27
{ "xMax": 1645, "xMin": 1636, "yMax": 523, "yMin": 487 }
person
[ 252, 30, 11, 255 ]
0
{ "xMax": 1429, "xMin": 1168, "yMax": 1023, "yMin": 14 }
person
[ 183, 38, 11, 255 ]
83.21
{ "xMax": 1616, "xMin": 1604, "yMax": 518, "yMin": 484 }
End of preview.

UrbanSyn Dataset

UrbanSyn is an open synthetic dataset featuring photorealistic driving scenes. It contains ground-truth annotations for semantic segmentation, scene depth, panoptic instance segmentation, and 2-D bounding boxes. Website https://urbansyn.org

Overview

UrbanSyn is a diverse, compact, and photorealistic dataset that provides more than 7.5k synthetic annotated images. It was born to address the synth-to-real domain gap, contributing to unprecedented synthetic-only baselines used by domain adaptation (DA) methods.

- Reduce the synth-to-real domain gap

UrbanSyn dataset helps to reduce the domain gap by contributing to unprecedented synthetic-only baselines used by domain adaptation (DA) methods.

- Ground-truth annotations

UrbanSyn comes with photorealistic color images, per-pixel semantic segmentation, depth, instance panoptic segmentation, and 2-D bounding boxes.

- Open for research and commercial purposes

UrbanSyn may be used for research and commercial purposes. It is released publicly under the Creative Commons Attribution-Commercial-ShareAlike 4.0 license.

- High-degree of photorealism

UrbanSyn features highly realistic and curated driving scenarios leveraging procedurally-generated content and high-quality curated assets. To achieve UrbanSyn photorealism we leverage industry-standard unbiased path-tracing and AI-based denoising techniques.

White Paper

[Arxiv]

When using or referring to the UrbanSyn dataset in your research, please cite our white paper:

@misc{gomez2023one,
title={All for One, and One for All: UrbanSyn Dataset, the third Musketeer of Synthetic Driving Scenes},
author={Jose L. Gómez and Manuel Silva and Antonio Seoane and Agnès Borrás and Mario Noriega and Germán Ros and Jose A. Iglesias-Guitian and Antonio M. López},
year={2023},
eprint={2312.12176},
archivePrefix={arXiv},
primaryClass={cs.CV}
} 

Terms of Use

The UrbanSyn Dataset is provided by the Computer Vision Center (UAB) and CITIC (University of A Coruña).

UrbanSyn may be used for research and commercial purposes, and it is subject to the Creative Commons Attribution-Commercial-ShareAlike 4.0. A summary of the CC-BY-SA 4.0 licensing terms can be found [here].

Due to constraints from our asset providers for UrbanSyn, we prohibit the use of generative AI technologies for reverse engineering any assets or creating content for stock media platforms based on the UrbanSyn dataset.

While we strive to generate precise data, all information is presented 'as is' without any express or implied warranties. We explicitly disclaim all representations and warranties regarding the validity, scope, accuracy, completeness, safety, or utility of the licensed content, including any implied warranties of merchantability, fitness for a particular purpose, or otherwise.

Acknowledgements

Funded by Grant agreement PID2020-115734RB-C21 "SSL-ADA" and Grant agreement PID2020-115734RB-C22 "PGAS-ADA"

For more information about our team members and how to contact us, visit our website https://urbansyn.org

Folder structure and content

  • rgb: contains RGB images with a resolution of 2048x1024 in PNG format.
  • ss and ss_colour : contains the pixel-level semantic segmentation labels in grayscale (value = Class ID) and colour (value = Class RGB) respectively in PNG format. We follow the 19 training classes defined on Cityscapes:
    name trainId color
    'road' 0 (128, 64,128)
    'sidewalk' 1 (244, 35,232)
    'building' 2 ( 70, 70, 70)
    'wall' 3 (102,102,156)
    'fence' 4 (190,153,153)
    'pole' 5 (153,153,153)
    'traffic light' 6 (250,170, 30)
    'traffic sign' 7 (220,220, 0)
    'vegetation' 8 (107,142, 35)
    'terrain' 9 (152,251,152)
    'sky' 10 ( 70,130,180)
    'person' 11 (220, 20, 60)
    'rider' 12 (255, 0, 0)
    'car' 13 ( 0, 0,142)
    'truck' 14 ( 0, 0, 70)
    'bus' 15 ( 0, 60,100)
    'train' 16 ( 0, 80,100)
    'motorcycle' 17 ( 0, 0,230)
    'bicycle' 18 (119, 11, 32)
    'unlabeled' 19 ( 0, 0, 0)
  • panoptic: contains the instance segmentation of the dynamic objects of the image in PNG format. Each instance is codified using the RGB channels, where RG corresponds to the instance number and B to the class ID. Dynamic objects are Person, Rider, Car, Truck, Bus, Train, Motorcycle and Bicycle.
  • bbox2D: contains the 2D bounding boxes and Instances information for all the dynamic objects in the image up to 110 meters of distance from the camera and bigger than 150 pixels. We provide the annotations in a json file with the next structure:
    • bbox: provides the bounding box size determined by the top left corner (xMin, yMin) and Bottom right corner (xMax, YMax).
    • color: corresponds to the colour of the instance in the panoptic instance segmentation map inside panoptic folder.
    • label: defines the class name
    • occlusion_percentage: provides the occlusion percentatge of the object. Being 0 not occluded and 100 fully occluded.
  • depth: contains the depth map of the image in EXR format.

Download locally with huggingface_hub library

  • Install huggingface_hub library

  • You can download the dataset on Python this way:

    from huggingface_hub import snapshot_download

    snapshot_download(repo_id="UrbanSyn/UrbanSyn", repo_type="dataset")

  • More information about how to download and additional options can be found here

Downloads last month
61
Edit dataset card