update card
Browse files
README.md
CHANGED
@@ -11,16 +11,15 @@ model-index:
|
|
11 |
|
12 |
<!-- Provide a quick summary of what the model is/does. -->
|
13 |
|
14 |
-
[`PatchTST`](https://huggingface.co/docs/transformers/model_doc/patchtst) is a transformer-based model for time series modeling tasks, including forecasting, regression, and classification.
|
15 |
-
|
16 |
-
|
17 |
-
In this context, we offer a pre-trained `PatchTST` model encompassing all seven channels of the `ETTh1` dataset.
|
18 |
This particular pre-trained model produces a Mean Squared Error (MSE) of 0.3881 on the `test` split of the `ETTh1` dataset when forecasting 96 hours into the future with a historical data window of 512 hours.
|
19 |
|
20 |
-
For training and evaluating a `PatchTST` model, you can refer to
|
21 |
|
22 |
## Model Details
|
23 |
|
|
|
|
|
24 |
The `PatchTST` model was proposed in A Time Series is Worth [64 Words: Long-term Forecasting with Transformers](https://arxiv.org/abs/2211.14730) by Yuqi Nie, Nam H. Nguyen, Phanwadee Sinthong, Jayant Kalagnanam.
|
25 |
|
26 |
At a high level the model vectorizes time series into patches of a given size and encodes the resulting sequence of vectors via a Transformer that then outputs the prediction length forecast via an appropriate head.
|
@@ -29,12 +28,6 @@ The model is based on two key components: (i) segmentation of time series into s
|
|
29 |
|
30 |
In addition, PatchTST has a modular design to seamlessly support masked time series pre-training as well as direct time series forecasting, classification, and regression.
|
31 |
|
32 |
-
|
33 |
-
### Model Description
|
34 |
-
|
35 |
-
<!-- Provide a longer summary of what this model is. -->
|
36 |
-
|
37 |
-
|
38 |
<img src="patchtst_architecture.png" alt="Architecture" width="600" />
|
39 |
|
40 |
### Model Sources
|
|
|
11 |
|
12 |
<!-- Provide a quick summary of what the model is/does. -->
|
13 |
|
14 |
+
[`PatchTST`](https://huggingface.co/docs/transformers/model_doc/patchtst) is a transformer-based model for time series modeling tasks, including forecasting, regression, and classification. This repository contains a pre-trained `PatchTST` model encompassing all seven channels of the `ETTh1` dataset.
|
|
|
|
|
|
|
15 |
This particular pre-trained model produces a Mean Squared Error (MSE) of 0.3881 on the `test` split of the `ETTh1` dataset when forecasting 96 hours into the future with a historical data window of 512 hours.
|
16 |
|
17 |
+
For training and evaluating a `PatchTST` model, you can refer to this [demo notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/patch_tst_getting_started.ipynb).
|
18 |
|
19 |
## Model Details
|
20 |
|
21 |
+
### Model Description
|
22 |
+
|
23 |
The `PatchTST` model was proposed in A Time Series is Worth [64 Words: Long-term Forecasting with Transformers](https://arxiv.org/abs/2211.14730) by Yuqi Nie, Nam H. Nguyen, Phanwadee Sinthong, Jayant Kalagnanam.
|
24 |
|
25 |
At a high level the model vectorizes time series into patches of a given size and encodes the resulting sequence of vectors via a Transformer that then outputs the prediction length forecast via an appropriate head.
|
|
|
28 |
|
29 |
In addition, PatchTST has a modular design to seamlessly support masked time series pre-training as well as direct time series forecasting, classification, and regression.
|
30 |
|
|
|
|
|
|
|
|
|
|
|
|
|
31 |
<img src="patchtst_architecture.png" alt="Architecture" width="600" />
|
32 |
|
33 |
### Model Sources
|