| title | emoji | colorFrom | colorTo | sdk | app_file | pinned | license | short_description | sdk_version |
|---|---|---|---|---|---|---|---|---|---|
Time Series Benchmark |
🥇 |
green |
indigo |
gradio |
app.py |
true |
mit |
A Benchmark for Advanced Deep Time Series Models. |
5.19.0 |
Most of the variables to change for a default leaderboard are in src/env.py (replace the path for your leaderboard) and src/about.py (for tasks).
Results files should have the following format and be stored as json files:
{
"config": {
"model_dtype": "torch.float16", # or torch.bfloat16 or 8bit or 4bit
"model_name": "path of the model on the hub: org/model",
"model_sha": "revision on the hub",
},
"results": {
"task_name": {
"metric_name": score,
},
"task_name2": {
"metric_name": score,
}
}
}Request files are created automatically by this tool.
If you encounter problem on the space, don't hesitate to restart it to remove the create eval-queue, eval-queue-bk, eval-results and eval-results-bk created folder.
You'll find
- the main table' columns names and properties in
src/display/utils.py - the logic to read all results and request files, then convert them in dataframe lines, in
src/leaderboard/read_evals.py, andsrc/populate.py - the logic to allow or filter submissions in
src/submission/submit.pyandsrc/submission/check_validity.py
You can find APIs in app.py, we give init_leaderboard a leaderboard object and generate a corresponding table.
Like GIFT-EVAL, we can divide the leaderboard object into data_part dataframe and model_part dataframe.
model_part is easy to get, you can check GIFT-EVAL model_info_df and observe how it comes into being.
In our repo's app.py, there are three model_info_df defined by the same function, but we give it a different path(done in envs.py)
Now we first need to add model_configs and long_term_forceasting_results into all model folders.(from arxiv articles data)
Then we base on the format of results to write a function to turn it into a correct dataframe, like grouped_df function in utils.py in GIFT-EVAL. This can be done easier because our results is less and easy to understand.
-
Timer-XL, iTransformer, DLinear, PatchTST, TimesNet, Stationary, Autoformer from Timer-XL paper https://arxiv.org/abs/2410.04803 Table 11.
-
Crossformer, TimerXer from TimerXer paper https://arxiv.org/pdf/2402.19072 Table 12.
-
TimeMixer, Informer from TimeMixer paper https://arxiv.org/pdf/2405.14616 Table 13.
- All models from Timer-XL paper https://arxiv.org/abs/2410.04803 Table 12.
- All models from TimesBert paper https://arxiv.org/abs/2502.21245 Table 5.
The AVG column now is simply the unweighted average of all datasets.