wmgifford commited on
Commit
6b9d616
1 Parent(s): b995fed

update card

Browse files
Files changed (1) hide show
  1. README.md +4 -11
README.md CHANGED
@@ -11,16 +11,15 @@ model-index:
11
 
12
  <!-- Provide a quick summary of what the model is/does. -->
13
 
14
- [`PatchTST`](https://huggingface.co/docs/transformers/model_doc/patchtst) is a transformer-based model for time series modeling tasks, including forecasting, regression, and classification.
15
-
16
-
17
- In this context, we offer a pre-trained `PatchTST` model encompassing all seven channels of the `ETTh1` dataset.
18
  This particular pre-trained model produces a Mean Squared Error (MSE) of 0.3881 on the `test` split of the `ETTh1` dataset when forecasting 96 hours into the future with a historical data window of 512 hours.
19
 
20
- For training and evaluating a `PatchTST` model, you can refer to [this notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/patch_tst_getting_started.ipynb).
21
 
22
  ## Model Details
23
 
 
 
24
  The `PatchTST` model was proposed in A Time Series is Worth [64 Words: Long-term Forecasting with Transformers](https://arxiv.org/abs/2211.14730) by Yuqi Nie, Nam H. Nguyen, Phanwadee Sinthong, Jayant Kalagnanam.
25
 
26
  At a high level the model vectorizes time series into patches of a given size and encodes the resulting sequence of vectors via a Transformer that then outputs the prediction length forecast via an appropriate head.
@@ -29,12 +28,6 @@ The model is based on two key components: (i) segmentation of time series into s
29
 
30
  In addition, PatchTST has a modular design to seamlessly support masked time series pre-training as well as direct time series forecasting, classification, and regression.
31
 
32
-
33
- ### Model Description
34
-
35
- <!-- Provide a longer summary of what this model is. -->
36
-
37
-
38
  <img src="patchtst_architecture.png" alt="Architecture" width="600" />
39
 
40
  ### Model Sources
 
11
 
12
  <!-- Provide a quick summary of what the model is/does. -->
13
 
14
+ [`PatchTST`](https://huggingface.co/docs/transformers/model_doc/patchtst) is a transformer-based model for time series modeling tasks, including forecasting, regression, and classification. This repository contains a pre-trained `PatchTST` model encompassing all seven channels of the `ETTh1` dataset.
 
 
 
15
  This particular pre-trained model produces a Mean Squared Error (MSE) of 0.3881 on the `test` split of the `ETTh1` dataset when forecasting 96 hours into the future with a historical data window of 512 hours.
16
 
17
+ For training and evaluating a `PatchTST` model, you can refer to this [demo notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/patch_tst_getting_started.ipynb).
18
 
19
  ## Model Details
20
 
21
+ ### Model Description
22
+
23
  The `PatchTST` model was proposed in A Time Series is Worth [64 Words: Long-term Forecasting with Transformers](https://arxiv.org/abs/2211.14730) by Yuqi Nie, Nam H. Nguyen, Phanwadee Sinthong, Jayant Kalagnanam.
24
 
25
  At a high level the model vectorizes time series into patches of a given size and encodes the resulting sequence of vectors via a Transformer that then outputs the prediction length forecast via an appropriate head.
 
28
 
29
  In addition, PatchTST has a modular design to seamlessly support masked time series pre-training as well as direct time series forecasting, classification, and regression.
30
 
 
 
 
 
 
 
31
  <img src="patchtst_architecture.png" alt="Architecture" width="600" />
32
 
33
  ### Model Sources