|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""QuakeFlow_NC: A dataset of earthquake waveforms organized by earthquake events and based on the HDF5 format.""" |
|
|
|
|
|
import h5py |
|
import numpy as np |
|
import torch |
|
from typing import Dict, List, Optional, Tuple, Union |
|
|
|
import datasets |
|
|
|
|
|
|
|
|
|
_CITATION = """\ |
|
@InProceedings{huggingface:dataset, |
|
title = {NCEDC dataset for QuakeFlow}, |
|
author={Zhu et al.}, |
|
year={2023} |
|
} |
|
""" |
|
|
|
|
|
|
|
_DESCRIPTION = """\ |
|
A dataset of earthquake waveforms organized by earthquake events and based on the HDF5 format. |
|
""" |
|
|
|
|
|
_HOMEPAGE = "" |
|
|
|
|
|
_LICENSE = "" |
|
|
|
|
|
|
|
|
|
_REPO = "https://huggingface.co/datasets/AI4EPS/quakeflow_nc/resolve/main/data" |
|
_FILENAMES = ["NC1970-1989.h5", "NC1990-1994.h5", "NC1995-1999.h5", "NC2000-2004.h5", "NC2005-2009.h5", "NC2010.h5", "NC2011.h5", "NC2012.h5", "NC2013.h5", "NC2014.h5", "NC2015.h5", "NC2016.h5", "NC2017.h5", "NC2018.h5", "NC2019.h5", "NC2020.h5"] |
|
_URLS = { |
|
"station": [f"{_REPO}/{x}" for x in _FILENAMES], |
|
"event": [f"{_REPO}/{x}" for x in _FILENAMES], |
|
} |
|
|
|
class BatchBuilderConfig(datasets.BuilderConfig): |
|
""" |
|
yield a batch of event-based sample, so the number of sample stations can vary among batches |
|
Batch Config for QuakeFlow_NC |
|
:param batch_size: number of samples in a batch |
|
:param num_stations_list: possible number of stations in a batch |
|
""" |
|
def __init__(self, batch_size: int, num_stations_list: List, **kwargs): |
|
super().__init__(**kwargs) |
|
self.batch_size = batch_size |
|
self.num_stations_list = num_stations_list |
|
|
|
|
|
|
|
class QuakeFlow_NC(datasets.GeneratorBasedBuilder): |
|
"""QuakeFlow_NC: A dataset of earthquake waveforms organized by earthquake events and based on the HDF5 format.""" |
|
|
|
VERSION = datasets.Version("1.1.0") |
|
|
|
degree2km = 111.32 |
|
nt = 8192 |
|
feature_nt = 512 |
|
feature_scale = int(nt / feature_nt) |
|
sampling_rate=100.0 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
BUILDER_CONFIGS = [ |
|
datasets.BuilderConfig(name="station", version=VERSION, description="yield station-based samples one by one"), |
|
datasets.BuilderConfig(name="event", version=VERSION, description="yield event-based samples one by one"), |
|
] |
|
|
|
DEFAULT_CONFIG_NAME = "station" |
|
|
|
def _info(self): |
|
|
|
if self.config.name=="station": |
|
features=datasets.Features( |
|
{ |
|
"waveform": datasets.Array3D(shape=(3, self.nt), dtype='float32'), |
|
"phase_pick": datasets.Array3D(shape=(3, self.nt), dtype='float32'), |
|
"event_location": datasets.Sequence(datasets.Value("float32")), |
|
"station_location": datasets.Sequence(datasets.Value("float32")), |
|
}) |
|
|
|
elif self.config.name=="event": |
|
features=datasets.Features( |
|
{ |
|
"waveform": datasets.Array3D(shape=(None, 3, self.nt), dtype='float32'), |
|
"phase_pick": datasets.Array3D(shape=(None, 3, self.nt), dtype='float32'), |
|
"event_location": datasets.Sequence(datasets.Value("float32")), |
|
"station_location": datasets.Array2D(shape=(None, 3), dtype="float32"), |
|
} |
|
) |
|
|
|
return datasets.DatasetInfo( |
|
|
|
description=_DESCRIPTION, |
|
|
|
features=features, |
|
|
|
|
|
|
|
|
|
homepage=_HOMEPAGE, |
|
|
|
license=_LICENSE, |
|
|
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
|
|
|
|
|
|
|
|
|
|
|
|
urls = _URLS[self.config.name] |
|
|
|
files = dl_manager.download_and_extract(urls) |
|
|
|
|
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
|
|
gen_kwargs={ |
|
"filepath": files[:-1], |
|
"split": "train", |
|
}, |
|
), |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
|
|
gen_kwargs={ |
|
"filepath": files[-1:], |
|
"split": "test" |
|
}, |
|
), |
|
] |
|
|
|
|
|
|
|
|
|
def _generate_examples(self, filepath, split): |
|
|
|
|
|
|
|
for file in filepath: |
|
with h5py.File(file, "r") as fp: |
|
|
|
for event_id in fp.keys(): |
|
event = fp[event_id] |
|
station_ids = list(event.keys()) |
|
if self.config.name=="station": |
|
waveforms = np.zeros([3, self.nt], dtype="float32") |
|
phase_pick = np.zeros_like(waveforms) |
|
attrs = event.attrs |
|
event_location = [attrs["longitude"], attrs["latitude"], attrs["depth_km"], attrs["event_time_index"]] |
|
|
|
for i, sta_id in enumerate(station_ids): |
|
waveforms[:, :self.nt] = event[sta_id][:,:self.nt] |
|
attrs = event[sta_id].attrs |
|
p_picks = attrs["phase_index"][attrs["phase_type"] == "P"] |
|
s_picks = attrs["phase_index"][attrs["phase_type"] == "S"] |
|
phase_pick[:, :self.nt] = generate_label([p_picks, s_picks], nt=self.nt) |
|
station_location = [attrs["longitude"], attrs["latitude"], -attrs["elevation_m"]/1e3] |
|
|
|
yield f"{event_id}/{sta_id}", { |
|
"waveform": torch.from_numpy(waveforms).float(), |
|
"phase_pick": torch.from_numpy(phase_pick).float(), |
|
"event_location": torch.from_numpy(np.array(event_location)).float(), |
|
"station_location": torch.from_numpy(np.array(station_location)).float(), |
|
} |
|
|
|
elif self.config.name=="event": |
|
waveforms = np.zeros([len(station_ids), 3, self.nt], dtype="float32") |
|
phase_pick = np.zeros_like(waveforms) |
|
attrs = event.attrs |
|
event_location = [attrs["longitude"], attrs["latitude"], attrs["depth_km"], attrs["event_time_index"]] |
|
station_location = [] |
|
|
|
for i, sta_id in enumerate(station_ids): |
|
waveforms[i, :, :self.nt] = event[sta_id][:,:self.nt] |
|
attrs = event[sta_id].attrs |
|
p_picks = attrs["phase_index"][attrs["phase_type"] == "P"] |
|
s_picks = attrs["phase_index"][attrs["phase_type"] == "S"] |
|
phase_pick[i, :, :] = generate_label([p_picks, s_picks], nt=self.nt) |
|
station_location.append([attrs["longitude"], attrs["latitude"], -attrs["elevation_m"]/1e3]) |
|
|
|
yield event_id, { |
|
"waveform": torch.from_numpy(waveforms).float(), |
|
"phase_pick": torch.from_numpy(phase_pick).float(), |
|
"event_location": torch.from_numpy(np.array(event_location)).float(), |
|
"station_location": torch.from_numpy(np.array(station_location)).float(), |
|
} |
|
|
|
|
|
def generate_label(phase_list, label_width=[150, 150], nt=8192): |
|
|
|
target = np.zeros([len(phase_list) + 1, nt], dtype=np.float32) |
|
|
|
for i, (picks, w) in enumerate(zip(phase_list, label_width)): |
|
for phase_time in picks: |
|
t = np.arange(nt) - phase_time |
|
gaussian = np.exp(-(t**2) / (2 * (w / 6) ** 2)) |
|
gaussian[gaussian < 0.1] = 0.0 |
|
target[i + 1, :] += gaussian |
|
|
|
target[0:1, :] = np.maximum(0, 1 - np.sum(target[1:, :], axis=0, keepdims=True)) |
|
|
|
return target |