The viewer is disabled because this dataset repo requires arbitrary Python code execution. Please consider removing the loading script and relying on automated data support (you can use convert_to_parquet from the datasets library). If this is not possible, please open a discussion for direct help.

Super-resolution of Velocity Fields in Three-dimensional Fluid Dynamics

This dataset loader attempts to reproduce the data of Wang et al. (2024)'s experiments on Super-resolution of 3D Turbulence.

References:

  • Wang et al. (2024): "Discovering Symmetry Breaking in Physical Systems with Relaxed Group Convolution"

Usage

For a given configuration (e.g. large_50):

>>> ds = datasets.load_dataset("dl2-g32/jhtdb", name="large_50")
>>> ds
DatasetDict({
    train: Dataset({
        features: ['lrs', 'hr'],
        num_rows: 40
    })
    validation: Dataset({
        features: ['lrs', 'hr'],
        num_rows: 5
    })
    test: Dataset({
        features: ['lrs', 'hr'],
        num_rows: 5
    })
})

Each split contains the input lrs which corresponds on a sequence of low resolution samples from time t - ws/2, ..., t, ... ts + ws/2 (ws = window size) and hr corresponds to the high resolution sample at time t. All the parameters per data point are specified in the corresponding metadata_*.csv.

Specifically, for the default configuration, for each datapoint we have 3 low resolution samples and 1 high resolution sample. Each of the former have shapes (3, 16, 16, 16) and the latter has shape (3, 64, 64, 64).

Replication

This dataset is entirely generated by scripts/generate.py and each configuration is fully specified in their corresponding scripts/*.yaml.

Usage

python -m scripts.generate --config scripts/small_100.yaml --token edu.jhu.pha.turbulence.testing-201311

This will create two folders on datasets/jhtdb:

  1. A tmp folder that will store all samples accross runs to serve as a cache.
  2. The corresponding subset, small_50 for example. This folder will contain a metadata_*.csv and data *.zip for each split.

Note:

  • For the small variants, the default token is enough, but for the large variants a token has to be requested. More details here.
  • For reference, the large_100 takes ~15 minutes to generate for a total of ~300MB.
Downloads last month
76
Edit dataset card