Datasets:

Size:
n<1K
ArXiv:
DOI:
License:
DarthReca's picture
Update README.md
e3ee8ea verified
---
license: openrail
task_categories:
- image-segmentation
pretty_name: California Burned Areas
size_categories:
- n<1K
tags:
- climate
---
# California Burned Areas Dataset
**Working on adding more data**
## Dataset Description
- **Paper:** [Pre-Print](https://arxiv.org/abs/2401.11519) and [Version of Record](https://ieeexplore.ieee.org/document/10261881)
### Dataset Summary
This dataset contains images from Sentinel-2 satellites taken before and after a wildfire.
The ground truth masks are provided by the California Department of Forestry and Fire Protection and they are mapped on the images.
### Supported Tasks
The dataset is designed to do binary semantic segmentation of burned vs unburned areas.
## Dataset Structure
We opted to use HDF5 to grant better portability and lower file size than GeoTIFF.
### Dataset opening
Using the dataset library, you download only the pre-patched raw version for simplicity.
```python
from dataset import load_dataset
# There are two available configurations, "post-fire" and "pre-post-fire."
dataset = load_dataset("DarthReca/california_burned_areas", name="post-fire")
```
The dataset was compressed using `h5py` and BZip2 from `hdf5plugin`. **WARNING: `hdf5plugin` is necessary to extract data**.
### Data Instances
Each matrix has a shape of 5490x5490xC, where C is 12 for pre-fire and post-fire images, while it is 0 for binary masks.
Pre-patched version is provided with matrices of size 512x512xC, too. In this case, only mask with at least one positive pixel is present.
You can find two versions of the dataset: _raw_ (without any transformation) and _normalized_ (with data normalized in the range 0-255).
Our suggestion is to use the _raw_ version to have the possibility to apply any wanted pre-processing step.
### Data Fields
In each standard HDF5 file, you can find post-fire, pre-fire images, and binary masks. The file is structured in this way:
```bash
β”œβ”€β”€ foldn
β”‚ β”œβ”€β”€ uid0
β”‚ β”‚ β”œβ”€β”€ pre_fire
β”‚ β”‚ β”œβ”€β”€ post_fire
β”‚ β”‚ β”œβ”€β”€ mask
β”‚ β”œβ”€β”€ uid1
β”‚ β”œβ”€β”€ post_fire
β”‚ β”œβ”€β”€ mask
β”‚
β”œβ”€β”€ foldm
β”œβ”€β”€ uid2
β”‚ β”œβ”€β”€ post_fire
β”‚ β”œβ”€β”€ mask
β”œβ”€β”€ uid3
β”œβ”€β”€ pre_fire
β”œβ”€β”€ post_fire
β”œβ”€β”€ mask
...
```
where `foldn` and `foldm` are fold names and `uidn` is a unique identifier for the wildfire.
For the pre-patched version, the structure is:
```bash
root
|
|-- uid0_x: {post_fire, pre_fire, mask}
|
|-- uid0_y: {post_fire, pre_fire, mask}
|
|-- uid1_x: {post_fire, mask}
|
...
```
the fold name is stored as an attribute.
### Data Splits
There are 5 random splits whose names are: 0, 1, 2, 3, and 4.
### Source Data
Data are collected directly from Copernicus Open Access Hub through the API. The band files are aggregated into one single matrix.
## Additional Information
### Licensing Information
This work is under OpenRAIL license.
### Citation Information
If you plan to use this dataset in your work please give the credit to Sentinel-2 mission and the California Department of Forestry and Fire Protection and cite using this BibTex:
```
@ARTICLE{cabuar,
author={Cambrin, Daniele Rege and Colomba, Luca and Garza, Paolo},
journal={IEEE Geoscience and Remote Sensing Magazine},
title={CaBuAr: California burned areas dataset for delineation [Software and Data Sets]},
year={2023},
volume={11},
number={3},
pages={106-113},
doi={10.1109/MGRS.2023.3292467}
}
```