Datasets:
Tasks:
Time Series Forecasting
Multilinguality:
monolingual
Size Categories:
1K<n<10K
Language Creators:
found
Annotations Creators:
no-annotation
Source Datasets:
original
Tags:
License:
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor. | |
# | |
# Licensed under the Apache License, Version 2.0 (the "License"); | |
# you may not use this file except in compliance with the License. | |
# You may obtain a copy of the License at | |
# | |
# http://www.apache.org/licenses/LICENSE-2.0 | |
# | |
# Unless required by applicable law or agreed to in writing, software | |
# distributed under the License is distributed on an "AS IS" BASIS, | |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
# See the License for the specific language governing permissions and | |
# limitations under the License. | |
"""Monash Time Series Forecasting Repository Dataset.""" | |
from dataclasses import dataclass | |
from datetime import datetime | |
from pathlib import Path | |
from typing import List, Optional | |
import numpy as np | |
from pandas.tseries.frequencies import to_offset | |
import datasets | |
from .utils import convert_tsf_to_dataframe, frequency_converter | |
# Find for instance the citation on arxiv or on the dataset repo/website | |
_CITATION = """\ | |
@InProceedings{godahewa2021monash, | |
author = "Godahewa, Rakshitha and Bergmeir, Christoph and Webb, Geoffrey I. and Hyndman, Rob J. and Montero-Manso, Pablo", | |
title = "Monash Time Series Forecasting Archive", | |
booktitle = "Neural Information Processing Systems Track on Datasets and Benchmarks", | |
year = "2021", | |
note = "forthcoming" | |
} | |
""" | |
_DESCRIPTION = """\ | |
Monash Time Series Forecasting Repository which contains 30+ datasets of related time series for global forecasting research. This repository includes both real-world and competition time series datasets covering varied domains. | |
""" | |
_HOMEPAGE = "https://forecastingdata.org/" | |
_LICENSE = "The Creative Commons Attribution 4.0 International License. https://creativecommons.org/licenses/by/4.0/" | |
_ROOT_URL = "data" | |
class MonashTSFBuilderConfig(datasets.BuilderConfig): | |
"""MonashTSF builder config with some added meta data.""" | |
file_name: Optional[str] = None | |
url: Optional[str] = None | |
prediction_length: Optional[int] = None | |
item_id_column: Optional[str] = None | |
data_column: Optional[str] = None | |
target_fields: Optional[List[str]] = None | |
feat_dynamic_real_fields: Optional[List[str]] = None | |
multivariate: bool = False | |
rolling_evaluations: int = 1 | |
class MonashTSF(datasets.GeneratorBasedBuilder): | |
"""Builder of Monash Time Series Forecasting repository of datasets.""" | |
VERSION = datasets.Version("1.0.0") | |
BUILDER_CONFIG_CLASS = MonashTSFBuilderConfig | |
BUILDER_CONFIGS = [ | |
MonashTSFBuilderConfig( | |
name="weather", | |
version=VERSION, | |
description="3010 daily time series representing the variations of four weather variables: rain, mintemp, maxtemp and solar radiation, measured at the weather stations in Australia.", | |
url="https://zenodo.org/record/4654822", | |
file_name="weather_dataset.zip", | |
data_column="series_type", | |
), | |
MonashTSFBuilderConfig( | |
name="tourism_yearly", | |
version=VERSION, | |
description="This dataset contains 518 yearly time series used in the Kaggle Tourism forecasting competition.", | |
url="https://zenodo.org/record/4656103", | |
file_name="tourism_yearly_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="tourism_quarterly", | |
version=VERSION, | |
description="This dataset contains 427 quarterly time series used in the Kaggle Tourism forecasting competition.", | |
url="https://zenodo.org/record/4656093", | |
file_name="tourism_quarterly_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="tourism_monthly", | |
version=VERSION, | |
description="This dataset contains 366 monthly time series used in the Kaggle Tourism forecasting competition.", | |
url="https://zenodo.org/record/4656096", | |
file_name="tourism_monthly_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="cif_2016", | |
version=VERSION, | |
description="72 monthly time series originated from the banking domain used in the CIF 2016 forecasting competition.", | |
url="https://zenodo.org/record/4656042", | |
file_name="cif_2016_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="london_smart_meters", | |
version=VERSION, | |
description="5560 half hourly time series that represent the energy consumption readings of London households in kilowatt hour (kWh) from November 2011 to February 2014.", | |
url="https://zenodo.org/record/4656072", | |
file_name="london_smart_meters_dataset_with_missing_values.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="australian_electricity_demand", | |
version=VERSION, | |
description="5 time series representing the half hourly electricity demand of 5 states in Australia: Victoria, New South Wales, Queensland, Tasmania and South Australia.", | |
url="https://zenodo.org/record/4659727", | |
file_name="australian_electricity_demand_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="wind_farms_minutely", | |
version=VERSION, | |
description="Minutely time series representing the wind power production of 339 wind farms in Australia.", | |
url="https://zenodo.org/record/4654909", | |
file_name="wind_farms_minutely_dataset_with_missing_values.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="bitcoin", | |
version=VERSION, | |
description="18 daily time series including hash rate, block size, mining difficulty etc. as well as public opinion in the form of tweets and google searches mentioning the keyword bitcoin as potential influencer of the bitcoin price.", | |
url="https://zenodo.org/record/5121965", | |
file_name="bitcoin_dataset_with_missing_values.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="pedestrian_counts", | |
version=VERSION, | |
description="Hourly pedestrian counts captured from 66 sensors in Melbourne city starting from May 2009.", | |
url="https://zenodo.org/record/4656626", | |
file_name="pedestrian_counts_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="vehicle_trips", | |
version=VERSION, | |
description="329 daily time series representing the number of trips and vehicles belonging to a set of for-hire vehicle (FHV) companies.", | |
url="https://zenodo.org/record/5122535", | |
file_name="vehicle_trips_dataset_with_missing_values.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="kdd_cup_2018", | |
version=VERSION, | |
description="Hourly time series representing the air quality levels in 59 stations in 2 cities from 01/01/2017 to 31/03/2018.", | |
url="https://zenodo.org/record/4656719", | |
file_name="kdd_cup_2018_dataset_with_missing_values.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="nn5_daily", | |
version=VERSION, | |
description="111 time series to predicting the daily cash withdrawals from ATMs in UK.", | |
url="https://zenodo.org/record/4656110", | |
file_name="nn5_daily_dataset_with_missing_values.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="nn5_weekly", | |
version=VERSION, | |
description="111 time series to predicting the weekly cash withdrawals from ATMs in UK.", | |
url="https://zenodo.org/record/4656125", | |
file_name="nn5_weekly_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="kaggle_web_traffic", | |
version=VERSION, | |
description="145063 daily time series representing the number of hits or web traffic for a set of Wikipedia pages from 2015-07-01 to 2017-09-10.", | |
url="https://zenodo.org/record/4656080", | |
file_name="kaggle_web_traffic_dataset_with_missing_values.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="kaggle_web_traffic_weekly", | |
version=VERSION, | |
description="145063 daily time series representing the number of hits or web traffic for a set of Wikipedia pages from 2015-07-01 to 2017-09-10.", | |
url="https://zenodo.org/record/4656664", | |
file_name="kaggle_web_traffic_weekly_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="solar_10_minutes", | |
version=VERSION, | |
description="137 time series representing the solar power production recorded per every 10 minutes in Alabama state in 2006.", | |
url="https://zenodo.org/record/4656144", | |
file_name="solar_10_minutes_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="solar_weekly", | |
version=VERSION, | |
description="137 time series representing the weekly solar power production in Alabama state in 2006.", | |
url="https://zenodo.org/record/4656151", | |
file_name="solar_weekly_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="car_parts", | |
version=VERSION, | |
description="2674 intermittent monthly time series that represent car parts sales from January 1998 to March 2002.", | |
url="https://zenodo.org/record/4656022", | |
file_name="car_parts_dataset_with_missing_values.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="fred_md", | |
version=VERSION, | |
description="107 monthly time series showing a set of macro-economic indicators from the Federal Reserve Bank.", | |
url="https://zenodo.org/record/4654833", | |
file_name="fred_md_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="traffic_hourly", | |
version=VERSION, | |
description="862 hourly time series showing the road occupancy rates on the San Francisco Bay area freeways from 2015 to 2016.", | |
url="https://zenodo.org/record/4656132", | |
file_name="traffic_hourly_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="traffic_weekly", | |
version=VERSION, | |
description="862 weekly time series showing the road occupancy rates on the San Francisco Bay area freeways from 2015 to 2016.", | |
url="https://zenodo.org/record/4656135", | |
file_name="traffic_weekly_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="hospital", | |
version=VERSION, | |
description="767 monthly time series that represent the patient counts related to medical products from January 2000 to December 2006.", | |
url="https://zenodo.org/record/4656014", | |
file_name="hospital_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="covid_deaths", | |
version=VERSION, | |
description="266 daily time series that represent the COVID-19 deaths in a set of countries and states from 22/01/2020 to 20/08/2020.", | |
url="https://zenodo.org/record/4656009", | |
file_name="covid_deaths_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="sunspot", | |
version=VERSION, | |
description="A single very long daily time series of sunspot numbers from 1818-01-08 to 2020-05-31.", | |
url="https://zenodo.org/record/4654773", | |
file_name="sunspot_dataset_with_missing_values.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="saugeenday", | |
version=VERSION, | |
description="A single very long time series representing the daily mean flow of the Saugeen River at Walkerton in cubic meters per second from 01/01/1915 to 31/12/1979.", | |
url="https://zenodo.org/record/4656058", | |
file_name="saugeenday_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="us_births", | |
version=VERSION, | |
description="A single very long daily time series representing the number of births in US from 01/01/1969 to 31/12/1988.", | |
url="https://zenodo.org/record/4656049", | |
file_name="us_births_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="solar_4_seconds", | |
version=VERSION, | |
description="A single very long daily time series representing the solar power production in MW recorded per every 4 seconds starting from 01/08/2019.", | |
url="https://zenodo.org/record/4656027", | |
file_name="solar_4_seconds_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="wind_4_seconds", | |
version=VERSION, | |
description="A single very long daily time series representing the wind power production in MW recorded per every 4 seconds starting from 01/08/2019.", | |
url="https://zenodo.org/record/4656032", | |
file_name="wind_4_seconds_dataset.zip", | |
), | |
MonashTSFBuilderConfig( | |
name="rideshare", | |
version=VERSION, | |
description="156 hourly time series representations of attributes related to Uber and Lyft rideshare services for various locations in New York between 26/11/2018 and 18/12/2018.", | |
url="https://zenodo.org/record/5122114", | |
file_name="rideshare_dataset_with_missing_values.zip", | |
item_id_column=["source_location", "provider_name", "provider_service"], | |
data_column="type", | |
target_fields=[ | |
"price_min", | |
"price_mean", | |
"price_max", | |
"distance_min", | |
"distance_mean", | |
"distance_max", | |
"surge_min", | |
"surge_mean", | |
"surge_max", | |
"api_calls", | |
], | |
feat_dynamic_real_fields=["temp", "rain", "humidity", "clouds", "wind"], | |
multivariate=True, | |
), | |
MonashTSFBuilderConfig( | |
name="oikolab_weather", | |
version=VERSION, | |
description="Eight time series representing the hourly climate data nearby Monash University, Clayton, Victoria, Australia from 2010-01-01 to 2021-05-31", | |
url="https://zenodo.org/record/5184708", | |
file_name="oikolab_weather_dataset.zip", | |
data_column="type", | |
), | |
MonashTSFBuilderConfig( | |
name="temperature_rain", | |
version=VERSION, | |
description="32072 daily time series showing the temperature observations and rain forecasts, gathered by the Australian Bureau of Meteorology for 422 weather stations across Australia, between 02/05/2015 and 26/04/2017", | |
url="https://zenodo.org/record/5129073", | |
file_name="temperature_rain_dataset_with_missing_values.zip", | |
item_id_column="station_id", | |
data_column="obs_or_fcst", | |
target_fields=[ | |
"fcst_0_DailyPoP", | |
"fcst_0_DailyPoP1", | |
"fcst_0_DailyPoP10", | |
"fcst_0_DailyPoP15", | |
"fcst_0_DailyPoP25", | |
"fcst_0_DailyPoP5", | |
"fcst_0_DailyPoP50", | |
"fcst_0_DailyPrecip", | |
"fcst_0_DailyPrecip10Pct", | |
"fcst_0_DailyPrecip25Pct", | |
"fcst_0_DailyPrecip50Pct", | |
"fcst_0_DailyPrecip75Pct", | |
"fcst_1_DailyPoP", | |
"fcst_1_DailyPoP1", | |
"fcst_1_DailyPoP10", | |
"fcst_1_DailyPoP15", | |
"fcst_1_DailyPoP25", | |
"fcst_1_DailyPoP5", | |
"fcst_1_DailyPoP50", | |
"fcst_1_DailyPrecip", | |
"fcst_1_DailyPrecip10Pct", | |
"fcst_1_DailyPrecip25Pct", | |
"fcst_1_DailyPrecip50Pct", | |
"fcst_1_DailyPrecip75Pct", | |
"fcst_2_DailyPoP", | |
"fcst_2_DailyPoP1", | |
"fcst_2_DailyPoP10", | |
"fcst_2_DailyPoP15", | |
"fcst_2_DailyPoP25", | |
"fcst_2_DailyPoP5", | |
"fcst_2_DailyPoP50", | |
"fcst_2_DailyPrecip", | |
"fcst_2_DailyPrecip10Pct", | |
"fcst_2_DailyPrecip25Pct", | |
"fcst_2_DailyPrecip50Pct", | |
"fcst_2_DailyPrecip75Pct", | |
"fcst_3_DailyPoP", | |
"fcst_3_DailyPoP1", | |
"fcst_3_DailyPoP10", | |
"fcst_3_DailyPoP15", | |
"fcst_3_DailyPoP25", | |
"fcst_3_DailyPoP5", | |
"fcst_3_DailyPoP50", | |
"fcst_3_DailyPrecip", | |
"fcst_3_DailyPrecip10Pct", | |
"fcst_3_DailyPrecip25Pct", | |
"fcst_3_DailyPrecip50Pct", | |
"fcst_3_DailyPrecip75Pct", | |
"fcst_4_DailyPoP", | |
"fcst_4_DailyPoP1", | |
"fcst_4_DailyPoP10", | |
"fcst_4_DailyPoP15", | |
"fcst_4_DailyPoP25", | |
"fcst_4_DailyPoP5", | |
"fcst_4_DailyPoP50", | |
"fcst_4_DailyPrecip", | |
"fcst_4_DailyPrecip10Pct", | |
"fcst_4_DailyPrecip25Pct", | |
"fcst_4_DailyPrecip50Pct", | |
"fcst_4_DailyPrecip75Pct", | |
"fcst_5_DailyPoP", | |
"fcst_5_DailyPoP1", | |
"fcst_5_DailyPoP10", | |
"fcst_5_DailyPoP15", | |
"fcst_5_DailyPoP25", | |
"fcst_5_DailyPoP5", | |
"fcst_5_DailyPoP50", | |
"fcst_5_DailyPrecip", | |
"fcst_5_DailyPrecip10Pct", | |
"fcst_5_DailyPrecip25Pct", | |
"fcst_5_DailyPrecip50Pct", | |
"fcst_5_DailyPrecip75Pct", | |
], | |
feat_dynamic_real_fields=["T_MEAN", "PRCP_SUM", "T_MAX", "T_MIN"], | |
multivariate=True, | |
), | |
] | |
def _info(self): | |
if self.config.multivariate: | |
features = datasets.Features( | |
{ | |
"start": datasets.Value("timestamp[s]"), | |
"target": datasets.Sequence(datasets.Sequence(datasets.Value("float32"))), | |
"feat_static_cat": datasets.Sequence(datasets.Value("uint64")), | |
# "feat_static_real": datasets.Sequence(datasets.Value("float32")), | |
"feat_dynamic_real": datasets.Sequence(datasets.Sequence(datasets.Value("float32"))), | |
# "feat_dynamic_cat": datasets.Sequence(datasets.Sequence(datasets.Value("uint64"))), | |
"item_id": datasets.Value("string"), | |
} | |
) | |
else: | |
features = datasets.Features( | |
{ | |
"start": datasets.Value("timestamp[s]"), | |
"target": datasets.Sequence(datasets.Value("float32")), | |
"feat_static_cat": datasets.Sequence(datasets.Value("uint64")), | |
# "feat_static_real": datasets.Sequence(datasets.Value("float32")), | |
"feat_dynamic_real": datasets.Sequence(datasets.Sequence(datasets.Value("float32"))), | |
# "feat_dynamic_cat": datasets.Sequence(datasets.Sequence(datasets.Value("uint64"))), | |
"item_id": datasets.Value("string"), | |
} | |
) | |
return datasets.DatasetInfo( | |
description=_DESCRIPTION, | |
features=features, | |
homepage=_HOMEPAGE, | |
license=_LICENSE, | |
citation=_CITATION, | |
) | |
def _split_generators(self, dl_manager): | |
# dl_manager is a datasets.download.DownloadManager that can be used to download and extract URLS | |
# It can accept any type or nested list/dict and will give back the same structure with the url replaced with path to local files. | |
# By default the archives will be extracted and a path to a cached folder where they are extracted is returned instead of the archive | |
urls = f"{_ROOT_URL}/{self.config.file_name}" | |
data_dir = dl_manager.download_and_extract(urls) | |
file_path = Path(data_dir) / (self.config.file_name.split(".")[0] + ".tsf") | |
return [ | |
datasets.SplitGenerator( | |
name=datasets.Split.TRAIN, | |
# These kwargs will be passed to _generate_examples | |
gen_kwargs={ | |
"filepath": file_path, | |
"split": "train", | |
}, | |
), | |
datasets.SplitGenerator( | |
name=datasets.Split.TEST, | |
# These kwargs will be passed to _generate_examples | |
gen_kwargs={"filepath": file_path, "split": "test"}, | |
), | |
datasets.SplitGenerator( | |
name=datasets.Split.VALIDATION, | |
# These kwargs will be passed to _generate_examples | |
gen_kwargs={ | |
"filepath": file_path, | |
"split": "val", | |
}, | |
), | |
] | |
# method parameters are unpacked from `gen_kwargs` as given in `_split_generators` | |
def _generate_examples(self, filepath, split): | |
( | |
loaded_data, | |
frequency, | |
forecast_horizon, | |
_, | |
_, | |
) = convert_tsf_to_dataframe(filepath, value_column_name="target") | |
if forecast_horizon is None: | |
prediction_length_map = { | |
"S": 60, | |
"T": 60, | |
"H": 48, | |
"D": 30, | |
"W": 8, | |
"M": 12, | |
"Y": 4, | |
} | |
freq = frequency_converter(frequency) | |
freq = to_offset(freq).name | |
forecast_horizon = prediction_length_map[freq] | |
if self.config.prediction_length is not None: | |
forecast_horizon = self.config.prediction_length | |
if self.config.item_id_column is not None: | |
loaded_data.set_index(self.config.item_id_column, inplace=True) | |
loaded_data.sort_index(inplace=True) | |
for cat, item_id in enumerate(loaded_data.index.unique()): | |
ts = loaded_data.loc[item_id] | |
start = ts.start_timestamp[0] | |
if self.config.target_fields is not None: | |
target_fields = ts[ts[self.config.data_column].isin(self.config.target_fields)] | |
else: | |
target_fields = self.config.data_column.unique() | |
if self.config.feat_dynamic_real_fields is not None: | |
feat_dynamic_real_fields = ts[ | |
ts[self.config.data_column].isin(self.config.feat_dynamic_real_fields) | |
] | |
feat_dynamic_real = np.vstack(feat_dynamic_real_fields.target) | |
else: | |
feat_dynamic_real = None | |
target = np.vstack(target_fields.target) | |
feat_static_cat = [cat] | |
if split in ["train", "val"]: | |
offset = forecast_horizon * self.config.rolling_evaluations + forecast_horizon * (split == "train") | |
target = target[..., :-offset] | |
if self.config.feat_dynamic_real_fields is not None: | |
feat_dynamic_real = feat_dynamic_real[..., :-offset] | |
yield cat, { | |
"start": start, | |
"target": target, | |
"feat_dynamic_real": feat_dynamic_real, | |
"feat_static_cat": feat_static_cat, | |
"item_id": item_id, | |
} | |
else: | |
if self.config.target_fields is not None: | |
target_fields = loaded_data[loaded_data[self.config.data_column].isin(self.config.target_fields)] | |
else: | |
target_fields = loaded_data | |
if self.config.feat_dynamic_real_fields is not None: | |
feat_dynamic_real_fields = loaded_data[ | |
loaded_data[self.config.data_column].isin(self.config.feat_dynamic_real_fields) | |
] | |
else: | |
feat_dynamic_real_fields = None | |
for cat, ts in target_fields.iterrows(): | |
start = ts.get("start_timestamp", datetime.strptime("1900-01-01 00-00-00", "%Y-%m-%d %H-%M-%S")) | |
target = ts.target | |
if feat_dynamic_real_fields is not None: | |
feat_dynamic_real = np.vstack(feat_dynamic_real_fields.target) | |
else: | |
feat_dynamic_real = None | |
feat_static_cat = [cat] | |
if self.config.data_column is not None: | |
item_id = f"{ts.series_name}-{ts[self.config.data_column]}" | |
else: | |
item_id = ts.series_name | |
if split in ["train", "val"]: | |
offset = forecast_horizon * self.config.rolling_evaluations + forecast_horizon * (split == "train") | |
target = target[..., :-offset] | |
if feat_dynamic_real is not None: | |
feat_dynamic_real = feat_dynamic_real[..., :-offset] | |
yield cat, { | |
"start": start, | |
"target": target, | |
"feat_dynamic_real": feat_dynamic_real, | |
"feat_static_cat": feat_static_cat, | |
"item_id": item_id, | |
} | |