Chronos-Bolt is a family of pretrained time series forecasting models which can be used for zero-shot forecasting. It is based on the
T5 encoder-decoder architecture
and has been trained on nearly 100 billion time series observations. It chunks the historical time series context into patches of multiple observations, which are then input into the encoder. The decoder then uses these representations to directly generate quantile forecasts across multiple future steps—a method known as direct multi-step forecasting. Chronos-Bolt models are up to 250 times faster and 20 times more memory-efficient than the
original Chronos
models of the same size.
The following plot compares the inference time of Chronos-Bolt against the original Chronos models for forecasting 1024 time series with a context length of 512 observations and a prediction horizon of 64 steps.
Chronos-Bolt models are not only significantly faster but also more accurate than the original Chronos models. The following plot reports the probabilistic and point forecasting performance of Chronos-Bolt in terms of the
Weighted Quantile Loss (WQL)
and the
Mean Absolute Scaled Error (MASE)
, respectively, aggregated over 27 datasets (see the
Chronos paper
for details on this benchmark). Remarkably, despite having no prior exposure to these datasets during training, the zero-shot Chronos-Bolt models outperform commonly used statistical models and deep learning models that have been trained on these datasets (highlighted by *). Furthermore, they also perform better than other FMs, denoted by a +, which indicates that these models were pretrained on certain datasets in our benchmark and are not entirely zero-shot. Notably, Chronos-Bolt (Base) also surpasses the original Chronos (Large) model in terms of the forecasting accuracy while being over 600 times faster.
Chronos-Bolt models are available in the following sizes.
The recommended way of using Chronos for production use cases is through
AutoGluon
, which features effortless fine-tuning, augmenting Chronos models with exogenous information through covariate regressors, ensembling with other statistical and machine learning models, as well as seamless deployments on AWS with SageMaker.
Check out the AutoGluon Chronos
tutorial
.
A minimal example showing how to perform zero-shot inference using Chronos-Bolt with AutoGluon:
Alternatively, you can install the package in the GitHub
companion repo
.
This is intended for research purposes and provides a minimal interface to Chronos models.
Install the library by running:
pip install chronos-forecasting
A minimal example showing how to perform inference using Chronos-Bolt models:
import pandas as pd # requires: pip install pandasimport torch
from chronos import BaseChronosPipeline
pipeline = BaseChronosPipeline.from_pretrained(
"amazon/chronos-bolt-tiny",
device_map="cuda", # use "cpu" for CPU inference and "mps" for Apple Silicon
torch_dtype=torch.bfloat16,
)
df = pd.read_csv(
"https://raw.githubusercontent.com/AileenNielsen/TimeSeriesAnalysisWithPython/master/data/AirPassengers.csv"
)
# context must be either a 1D tensor, a list of 1D tensors,# or a left-padded 2D tensor with batch as the first dimension# Chronos-Bolt models generate quantile forecasts, so forecast has shape# [num_series, num_quantiles, prediction_length].
forecast = pipeline.predict(
context=torch.tensor(df["#Passengers"]), prediction_length=12
)
Citation
If you find Chronos or Chronos-Bolt models useful for your research, please consider citing the associated
paper
:
@article{ansari2024chronos,
title={Chronos: Learning the Language of Time Series},
author={Ansari, Abdul Fatir and Stella, Lorenzo and Turkmen, Caner and Zhang, Xiyuan, and Mercado, Pedro and Shen, Huibin and Shchur, Oleksandr and Rangapuram, Syama Syndar and Pineda Arango, Sebastian and Kapoor, Shubham and Zschiegner, Jasper and Maddix, Danielle C. and Mahoney, Michael W. and Torkkola, Kari and Gordon Wilson, Andrew and Bohlke-Schneider, Michael and Wang, Yuyang},
journal={Transactions on Machine Learning Research},
issn={2835-8856},
year={2024},
url={https://openreview.net/forum?id=gerNCVqqtR}
}
License
This project is licensed under the Apache-2.0 License.
Runs of anchovy chronos-bolt-tiny on huggingface.co
15
Total runs
0
24-hour runs
0
3-day runs
10
7-day runs
15
30-day runs
More Information About chronos-bolt-tiny huggingface.co Model
chronos-bolt-tiny huggingface.co is an AI model on huggingface.co that provides chronos-bolt-tiny's model effect (), which can be used instantly with this anchovy chronos-bolt-tiny model. huggingface.co supports a free trial of the chronos-bolt-tiny model, and also provides paid use of the chronos-bolt-tiny. Support call chronos-bolt-tiny model through api, including Node.js, Python, http.
chronos-bolt-tiny huggingface.co is an online trial and call api platform, which integrates chronos-bolt-tiny's modeling effects, including api services, and provides a free online trial of chronos-bolt-tiny, you can try chronos-bolt-tiny online for free by clicking the link below.
anchovy chronos-bolt-tiny online free url in huggingface.co:
chronos-bolt-tiny is an open source model from GitHub that offers a free installation service, and any user can find chronos-bolt-tiny on GitHub to install. At the same time, huggingface.co provides the effect of chronos-bolt-tiny install, users can directly use chronos-bolt-tiny installed effect in huggingface.co for debugging and trial. It also supports api for free installation.