Datasets:

Modalities:
Image
Languages:
English
DOI:
Libraries:
Datasets
License:
hls_burn_scars / hls_burn_scars.py
xhagrg's picture
Use without rasterio.
1864285
raw
history blame
3.05 kB
import os
from glob import glob
import datasets
_CITATION = """\
@software{HLS_Foundation_2023,
author = {Phillips, Christopher and Roy, Sujit and Ankur, Kumar and Ramachandran, Rahul},
doi = {10.57967/hf/0956},
month = aug,
title = {{HLS Foundation Burnscars Dataset}},
url = {https://huggingface.co/ibm-nasa-geospatial/hls_burn_scars},
year = {2023}
}
"""
_DESCRIPTION = """\
This dataset contains Harmonized Landsat and Sentinel-2 imagery of burn scars and the associated masks for the years 2018-2021 over the contiguous United States. There are 804 512x512 scenes. Its primary purpose is for training geospatial machine learning models.
"""
_HOMEPAGE = "https://huggingface.co/datasets/ibm-nasa-geospatial/hls_burn_scars"
_LICENSE = "cc-by-4.0"
_URLS = {
"hls_burn_scars": {
"train/val": "https://huggingface.co/datasets/ibm-nasa-geospatial/hls_burn_scars/resolve/main/hls_burn_scars.tar.gz"
}
}
class HLSBurnScars(datasets.GeneratorBasedBuilder):
"""MIT Scene Parsing Benchmark dataset."""
VERSION = datasets.Version("0.0.1")
BUILDER_CONFIGS = [
datasets.BuilderConfig(name="hls_burn_scars", version=VERSION, description=_DESCRIPTION),
]
def _info(self):
features = datasets.Features(
{
"image": datasets.Image(),
"annotation": datasets.Image(),
}
)
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
homepage=_HOMEPAGE,
license=_LICENSE,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
urls = _URLS[self.config.name]
data_dirs = dl_manager.download_and_extract(urls)
train_data = os.path.join(data_dirs['train/val'], "training")
val_data = os.path.join(data_dirs['train/val'], "validation")
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={
"data": train_data,
"split": "training",
},
),
datasets.SplitGenerator(
name=datasets.Split.VALIDATION,
gen_kwargs={
"data": val_data,
"split": "validation",
},
),
datasets.SplitGenerator(
name=datasets.Split.TEST,
gen_kwargs={
"data": val_data,
"split": "testing",
},
)
]
def _generate_examples(self, data, split):
files = glob(f"{data}/*_merged.tif")
for idx, filename in enumerate(files):
if filename.endswith("_merged.tif"):
annotation_filename = filename.replace('_merged.tif', '.mask.tif')
yield idx, {
"image": {"path": filename},
"annotation": {"path": annotation_filename}
}