arjunashok commited on
Commit
f936da5
β€’
1 Parent(s): 3dd163d

Updates from GitHub README

Browse files
Files changed (1) hide show
  1. README.md +46 -12
README.md CHANGED
@@ -8,50 +8,84 @@ tags:
8
  - time series foundation models
9
  - time-series
10
  ---
 
 
11
  # Lag-Llama: Towards Foundation Models for Probabilistic Time Series Forecasting
12
 
13
  ![lag-llama-architecture](images/lagllama.webp)
14
 
15
  Lag-Llama is the <b>first open-source foundation model for time series forecasting</b>!
16
 
17
- [[Tweet Thread](https://twitter.com/arjunashok37/status/1755261111233114165)] [[Model Weights](https://huggingface.co/time-series-foundation-models/Lag-Llama)] [[Colab Demo on Zero-Shot Forecasting](https://colab.research.google.com/drive/13HHKYL_HflHBKxDWycXgIUAHSeHRR5eo?usp=sharing)] [[GitHub](https://github.com/time-series-foundation-models/lag-llama)] [[Paper](https://arxiv.org/abs/2310.08278)]
18
 
19
  ____
20
  This HuggingFace model houses the <a href="https://huggingface.co/time-series-foundation-models/Lag-Llama/blob/main/lag-llama.ckpt" target="_blank">pretrained checkpoint</a> of Lag-Llama.
21
 
22
  ____
23
 
24
- * **Coming Next**: Fine-tuning scripts with examples on real-world datasets and best practices in using Lag-Llama!πŸš€
25
-
26
  <b>Updates</b>:
27
 
28
- * **17-Feb-2024**: We have released a new updated [Colab Demo](https://colab.research.google.com/drive/1XxrLW9VGPlZDw3efTvUi0hQimgJOwQG6?usp=sharing) for zero-shot forecasting that shows how one can load time series of different formats.
 
 
 
29
  * **7-Feb-2024**: We released Lag-Llama, with open-source model checkpoints and a Colab Demo for zero-shot forecasting.
30
 
31
  ____
32
 
33
- <b>Current Features:</b>
 
 
 
 
 
 
34
 
35
- πŸ’« <b>Zero-shot forecasting</b> on a dataset of <b>any frequency</b> for <b>any prediction length</b>, using the <a href="https://colab.research.google.com/drive/13HHKYL_HflHBKxDWycXgIUAHSeHRR5eo?usp=sharing" target="_blank">Colab Demo.</a><br/>
36
 
37
  ____
38
 
39
  Coming Soon:
40
 
41
- ⭐ An <b>online gradio demo</b> where you can upload time series and get zero-shot predictions and perform finetuning.
42
 
43
- ⭐ Features for <b>finetuning</b> the foundation model
44
 
45
- ⭐ Features for <b>pretraining</b> Lag-Llama on your own large-scale data
46
 
47
- ⭐ Scripts to <b>reproduce</b> all results in the paper.
48
 
 
49
 
50
  ____
51
 
52
- Stay Tuned!πŸ¦™
53
 
54
- ____
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
55
 
56
  ## Citing this work
57
 
 
8
  - time series foundation models
9
  - time-series
10
  ---
11
+
12
+
13
  # Lag-Llama: Towards Foundation Models for Probabilistic Time Series Forecasting
14
 
15
  ![lag-llama-architecture](images/lagllama.webp)
16
 
17
  Lag-Llama is the <b>first open-source foundation model for time series forecasting</b>!
18
 
19
+ [[Tweet Thread](https://twitter.com/arjunashok37/status/1755261111233114165)] [[Model Weights](https://huggingface.co/time-series-foundation-models/Lag-Llama)] [[Colab Demo 1: Zero-Shot Forecasting](https://colab.research.google.com/drive/1DRAzLUPxsd-0r8b-o4nlyFXrjw_ZajJJ?usp=sharing)] [[Colab Demo 2: (Preliminary Finetuning)](https://colab.research.google.com/drive/1uvTmh-pe1zO5TeaaRVDdoEWJ5dFDI-pA?usp=sharing)] [[GitHub](https://github.com/time-series-foundation-models/lag-llama)] [[Paper](https://arxiv.org/abs/2310.08278)]
20
 
21
  ____
22
  This HuggingFace model houses the <a href="https://huggingface.co/time-series-foundation-models/Lag-Llama/blob/main/lag-llama.ckpt" target="_blank">pretrained checkpoint</a> of Lag-Llama.
23
 
24
  ____
25
 
 
 
26
  <b>Updates</b>:
27
 
28
+ * **5-Apr-2024**: Added a [section](https://colab.research.google.com/drive/1DRAzLUPxsd-0r8b-o4nlyFXrjw_ZajJJ?authuser=1#scrollTo=Mj9LXMpJ01d7&line=6&uniqifier=1) in Colab Demo 1 on the importance of tuning the context length for zero-shot forecasting. Added a [best practices section](https://github.com/time-series-foundation-models/lag-llama?tab=readme-ov-file#best-practices) in the README; added recommendations for finetuning. These recommendations will be demonstrated with an example in [Colab Demo 2](https://colab.research.google.com/drive/1uvTmh-pe1zO5TeaaRVDdoEWJ5dFDI-pA?usp=sharing) soon.
29
+ * **4-Apr-2024**: We have updated our requirements file with new versions of certain packages. Please update/recreate your environments if you have previously used the code locally.
30
+ * **7-Mar-2024**: We have released a preliminary [Colab Demo 2](https://colab.research.google.com/drive/1uvTmh-pe1zO5TeaaRVDdoEWJ5dFDI-pA?usp=sharing) for finetuning. Please note this is a preliminary tutorial. We recommend taking a look at the best practices if you are finetuning the model or using it for benchmarking.
31
+ * **17-Feb-2024**: We have released a new updated [Colab Demo 1](https://colab.research.google.com/drive/1DRAzLUPxsd-0r8b-o4nlyFXrjw_ZajJJ?usp=sharing) for zero-shot forecasting that shows how one can load time series of different formats.
32
  * **7-Feb-2024**: We released Lag-Llama, with open-source model checkpoints and a Colab Demo for zero-shot forecasting.
33
 
34
  ____
35
 
36
+ Current Features:
37
+
38
+ πŸ’« <b>Zero-shot forecasting</b> on a dataset of <b>any frequency</b> for <b>any prediction length</b>, using <a href="https://colab.research.google.com/drive/1DRAzLUPxsd-0r8b-o4nlyFXrjw_ZajJJ?usp=sharing" target="_blank">Colab Demo 1.</a><br/>
39
+
40
+ πŸ’« <b>Finetuning</b> on a dataset using [Colab Demo 2](https://colab.research.google.com/drive/1uvTmh-pe1zO5TeaaRVDdoEWJ5dFDI-pA?usp=sharing).
41
+
42
+ **Note**: Please see the [best practices section](https://github.com/time-series-foundation-models/lag-llama?tab=readme-ov-file#best-practices) when using the model for zero-shot prediction and finetuning.
43
 
 
44
 
45
  ____
46
 
47
  Coming Soon:
48
 
49
+ ⭐ Scripts to pretrain Lag-Llama on your own large-scale data
50
 
51
+ ⭐ Scripts to <b>reproduce</b> all results in the paper.
52
 
53
+ ____
54
 
55
+ We are currently looking for contributors for the following:
56
 
57
+ ⭐ An <b>online gradio demo</b> where you can upload time series and get zero-shot predictions and perform finetuning.
58
 
59
  ____
60
 
61
+ ## Best Practices
62
 
63
+ Here are some general tips in using Lag-Llama.
64
+ <!-- We recommend reading the [paper](https://arxiv.org/abs/2310.08278) for all details about the model. -->
65
+
66
+ ### General Information
67
+
68
+ * Lag-Llama is a **probabilistic** forecasting model trained to output a probability distribution for each timestep to be predicted. For your own specific use-case, we would recommend benchmarking the zero-shot performance of the model on your data first, and then finetuning if necessary. As we show in our paper, Lag-Llama has strong zero-shot capabilities, but performs best when finetuned. The more data you finetune on, the better. For specific tips on applying on model zero-shot or on finetuning, please refer to the sections below.
69
+
70
+ #### Zero-Shot Forecasting
71
+
72
+ * Importantly, we recommend trying different **context lengths** (starting from $32$ which it was trained on) and identifying what works best for your data. As we show in [this section of the zero-shot forecasting demo](https://colab.research.google.com/drive/1DRAzLUPxsd-0r8b-o4nlyFXrjw_ZajJJ?authuser=1#scrollTo=Mj9LXMpJ01d7&line=6&uniqifier=1), the model's zero-shot performance improves as the context length is increased, until a certain context length which may be specific to your data. Further, we recommend enabling RoPE scaling for the model to work well with context lengths larger than what it was trained on.
73
+
74
+ #### Fine-Tuning
75
+
76
+ If you are trying to **benchmark** the performance of the model under finetuning, or trying to obtain maximum performance from the model:
77
+
78
+ * We recommend tuning two important hyperparameters for each dataset that you finetune on: the **context length** (suggested values: $32$, $64$, $128$, $256$, $512$, $1024$) and the **learning rate** (suggested values: $10^{-2}$, $5 * 10^{-3}$, $10^{-3}$, $5 * 10^{-3}$, $1 * 10^{-4}$, $5 * 10^{-4}$).
79
+ * We also highly recommend using a validation split of your dataset to early stop your model, with an early stopping patience of 50 epochs.
80
+
81
+ ## Contact
82
+
83
+ We are dedicated to ensuring the reproducility of our results, and would be happy to help clarify questions about benchmarking our model or about the experiments in the paper.
84
+ The quickest way to reach us would be by email. Please email **both**:
85
+ 1. [Arjun Ashok](https://ashok-arjun.github.io/) - arjun [dot] ashok [at] servicenow [dot] com
86
+ 2. [Kashif Rasul](https://scholar.google.de/citations?user=cfIrwmAAAAAJ&hl=en) - kashif [dot] rasul [at] gmail [dot] com
87
+
88
+ If you have questions about the model usage (or) code (or) have specific errors (eg. using it with your own dataset), it would be best to create an issue in the GitHub repository.
89
 
90
  ## Citing this work
91