Update README.md
Browse files
README.md
CHANGED
@@ -53,29 +53,22 @@ TTMs that can cater to many common forecasting settings in practice.
|
|
53 |
Each pre-trained model will be released in a different branch name in this model card. Kindly access the required model using our
|
54 |
getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/ttm_getting_started.ipynb) mentioning the branch name.
|
55 |
|
56 |
-
|
57 |
-
## Model Releases (along with the branch name where the models are stored):
|
58 |
|
|
|
59 |
|
|
|
|
|
60 |
|
61 |
-
|
62 |
-
in future. (branch name: main)
|
63 |
-
|
64 |
-
- **1024-96-r2**: Given the last 1024 time-points (i.e. context length), this model can forecast up to the next 96 time-points (i.e. forecast length)
|
65 |
-
in future. (branch name: 1024-96-r2) [[Benchmarks]]
|
66 |
-
|
67 |
-
- **1536-96-r2**: Given the last 1536 time-points (i.e. context length), this model can forecast up to the next 96 time-points (i.e. forecast length)
|
68 |
-
in future. (branch name: 1536-96-r2)
|
69 |
|
70 |
-
-
|
71 |
-
`512-336-r2`, `1024-336-r2`, `1536-336-r2`, `512-720-r2`, `1024-720-r2`, `1536-720-r2`
|
72 |
|
73 |
-
-
|
74 |
|
75 |
-
-
|
76 |
-
but can provide any forecast lengths up to 720 in get_model() to get the required model.
|
77 |
|
78 |
-
|
79 |
|
80 |
|
81 |
|
|
|
53 |
Each pre-trained model will be released in a different branch name in this model card. Kindly access the required model using our
|
54 |
getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/ttm_getting_started.ipynb) mentioning the branch name.
|
55 |
|
56 |
+
## Model Releases:
|
|
|
57 |
|
58 |
+
Given the variety of models included, please use the [[get_model]](https://github.com/ibm-granite/granite-tsfm/blob/main/tsfm_public/toolkit/get_model.py) utility to automatically select the required model based on your input context length and forecast length requirement.
|
59 |
|
60 |
+
There are several models available in different branches of this model card. The naming scheme follows the following format:
|
61 |
+
`<context length>-<prediction length>-<frequency prefix tuning indicator>-<pretraining metric>-<release number>`
|
62 |
|
63 |
+
- context length: The historical data used as input to the TTM model.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
64 |
|
65 |
+
- prediction length: The number of time points predicted by model (i.e., the forecast length)
|
|
|
66 |
|
67 |
+
- frequency tuning indicator ("ft" or missing): "ft" is used to indicate use of frequency prefix tuning. When enabled an extra embedding vector indicating the frequency of the data is added to the input of the model. If missing, only the context window is used by the model.
|
68 |
|
69 |
+
- pretraining metric ("mae" or missing): MAE indicates pertaining with mean absolute error loss, while missing indicates using mean squared error.
|
|
|
70 |
|
71 |
+
- release number ("r2" or "r2.1"): Indicates the model release; the release indicates which data was used to train the model. See "training data" below for more details on the data included in the particular training datasets.
|
72 |
|
73 |
|
74 |
|