marcus0074 commited on
Commit
6c1678c
·
verified ·
1 Parent(s): 877fa8a

time series foundation model

Browse files
Files changed (5) hide show
  1. README.md +109 -0
  2. config.json +47 -0
  3. generation_config.json +4 -0
  4. model.safetensors +3 -0
  5. training_args.bin +3 -0
README.md ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - generated_from_trainer
4
+ model-index:
5
+ - name: output
6
+ results: []
7
+ ---
8
+
9
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
+ should probably proofread and complete it, then remove this comment. -->
11
+
12
+ # output
13
+
14
+ This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
15
+ It achieves the following results on the evaluation set:
16
+ - Loss: 0.1542
17
+
18
+ ## Model description
19
+
20
+ More information needed
21
+
22
+ ## Intended uses & limitations
23
+
24
+ More information needed
25
+
26
+ ## Training and evaluation data
27
+
28
+ More information needed
29
+
30
+ ## Training procedure
31
+
32
+ ### Training hyperparameters
33
+
34
+ The following hyperparameters were used during training:
35
+ - learning_rate: 0.001
36
+ - train_batch_size: 64
37
+ - eval_batch_size: 64
38
+ - seed: 42
39
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
+ - lr_scheduler_type: linear
41
+ - num_epochs: 100
42
+
43
+ ### Training results
44
+
45
+ | Training Loss | Epoch | Step | Validation Loss |
46
+ |:-------------:|:-----:|:-----:|:---------------:|
47
+ | 0.2823 | 1.0 | 384 | 0.1897 |
48
+ | 0.2134 | 2.0 | 768 | 0.1762 |
49
+ | 0.1961 | 3.0 | 1152 | 0.1714 |
50
+ | 0.1855 | 4.0 | 1536 | 0.1693 |
51
+ | 0.1777 | 5.0 | 1920 | 0.1738 |
52
+ | 0.1713 | 6.0 | 2304 | 0.1674 |
53
+ | 0.1664 | 7.0 | 2688 | 0.1681 |
54
+ | 0.1613 | 8.0 | 3072 | 0.1733 |
55
+ | 0.1578 | 9.0 | 3456 | 0.1698 |
56
+ | 0.1542 | 10.0 | 3840 | 0.1622 |
57
+ | 0.1505 | 11.0 | 4224 | 0.1666 |
58
+ | 0.1475 | 12.0 | 4608 | 0.1655 |
59
+ | 0.1451 | 13.0 | 4992 | 0.1651 |
60
+ | 0.1426 | 14.0 | 5376 | 0.1646 |
61
+ | 0.1409 | 15.0 | 5760 | 0.1618 |
62
+ | 0.1385 | 16.0 | 6144 | 0.1617 |
63
+ | 0.1366 | 17.0 | 6528 | 0.1591 |
64
+ | 0.1347 | 18.0 | 6912 | 0.1628 |
65
+ | 0.1325 | 19.0 | 7296 | 0.1598 |
66
+ | 0.1313 | 20.0 | 7680 | 0.1606 |
67
+ | 0.1295 | 21.0 | 8064 | 0.1573 |
68
+ | 0.1285 | 22.0 | 8448 | 0.1587 |
69
+ | 0.1276 | 23.0 | 8832 | 0.1639 |
70
+ | 0.1258 | 24.0 | 9216 | 0.1608 |
71
+ | 0.1244 | 25.0 | 9600 | 0.1599 |
72
+ | 0.1234 | 26.0 | 9984 | 0.1584 |
73
+ | 0.1225 | 27.0 | 10368 | 0.1604 |
74
+ | 0.1214 | 28.0 | 10752 | 0.1570 |
75
+ | 0.1207 | 29.0 | 11136 | 0.1575 |
76
+ | 0.1195 | 30.0 | 11520 | 0.1563 |
77
+ | 0.1186 | 31.0 | 11904 | 0.1602 |
78
+ | 0.1177 | 32.0 | 12288 | 0.1595 |
79
+ | 0.1167 | 33.0 | 12672 | 0.1582 |
80
+ | 0.1159 | 34.0 | 13056 | 0.1556 |
81
+ | 0.1149 | 35.0 | 13440 | 0.1564 |
82
+ | 0.114 | 36.0 | 13824 | 0.1567 |
83
+ | 0.1132 | 37.0 | 14208 | 0.1551 |
84
+ | 0.1125 | 38.0 | 14592 | 0.1560 |
85
+ | 0.1113 | 39.0 | 14976 | 0.1537 |
86
+ | 0.1114 | 40.0 | 15360 | 0.1518 |
87
+ | 0.1103 | 41.0 | 15744 | 0.1585 |
88
+ | 0.1098 | 42.0 | 16128 | 0.1552 |
89
+ | 0.1094 | 43.0 | 16512 | 0.1533 |
90
+ | 0.1087 | 44.0 | 16896 | 0.1542 |
91
+ | 0.1081 | 45.0 | 17280 | 0.1505 |
92
+ | 0.1085 | 46.0 | 17664 | 0.1535 |
93
+ | 0.1075 | 47.0 | 18048 | 0.1526 |
94
+ | 0.1069 | 48.0 | 18432 | 0.1521 |
95
+ | 0.1067 | 49.0 | 18816 | 0.1532 |
96
+ | 0.1063 | 50.0 | 19200 | 0.1522 |
97
+ | 0.1056 | 51.0 | 19584 | 0.1522 |
98
+ | 0.1048 | 52.0 | 19968 | 0.1538 |
99
+ | 0.1048 | 53.0 | 20352 | 0.1534 |
100
+ | 0.1051 | 54.0 | 20736 | 0.1519 |
101
+ | 0.1045 | 55.0 | 21120 | 0.1542 |
102
+
103
+
104
+ ### Framework versions
105
+
106
+ - Transformers 4.37.2
107
+ - Pytorch 2.2.0+cu121
108
+ - Datasets 2.17.0
109
+ - Tokenizers 0.15.2
config.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "PatchTSMixerForPrediction"
4
+ ],
5
+ "channel_consistent_masking": true,
6
+ "context_length": 336,
7
+ "d_model": 16,
8
+ "distribution_output": "student_t",
9
+ "dropout": 0.2,
10
+ "expansion_factor": 2,
11
+ "gated_attn": true,
12
+ "head_aggregation": "max_pool",
13
+ "head_dropout": 0.2,
14
+ "init_std": 0.02,
15
+ "loss": "mse",
16
+ "mask_type": "random",
17
+ "mask_value": 0,
18
+ "masked_loss": true,
19
+ "mode": "common_channel",
20
+ "model_type": "patchtsmixer",
21
+ "norm_eps": 1e-05,
22
+ "norm_mlp": "LayerNorm",
23
+ "num_forecast_mask_patches": [
24
+ 2
25
+ ],
26
+ "num_input_channels": 5,
27
+ "num_layers": 8,
28
+ "num_parallel_samples": 100,
29
+ "num_patches": 42,
30
+ "num_targets": 3,
31
+ "output_range": null,
32
+ "patch_last": true,
33
+ "patch_length": 8,
34
+ "patch_stride": 8,
35
+ "positional_encoding_type": "sincos",
36
+ "post_init": false,
37
+ "prediction_channel_indices": null,
38
+ "prediction_length": 24,
39
+ "random_mask_ratio": 0.5,
40
+ "scaling": "std",
41
+ "self_attn": false,
42
+ "self_attn_heads": 1,
43
+ "torch_dtype": "float32",
44
+ "transformers_version": "4.37.2",
45
+ "unmasked_channel_indices": null,
46
+ "use_positional_encoding": false
47
+ }
generation_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "transformers_version": "4.37.2"
4
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:766cbfc72ba5977581332849849ab68c83e73d1af16de5b1dcec5914699e29f5
3
+ size 415592
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e020b16425ffb87111456e7cacf5a0c4c58202d0204f4e43729cc6170830d253
3
+ size 4728