Google's 200M-parameter time-series foundation model with 16k context
Competing models and resources
- Commenters list several alternative time-series models and libraries: Datadog’s foundation model, Moment, TabPFN, OpenTSLM, Nixtla, Prophet, Amazon’s Chronos, and models on Salesforce’s GIFT leaderboard.
- Some see this as an emerging “foundation model” space for time series with multiple active contenders.
Architecture, training, and scale
- Links are shared to Google’s blog and the full paper.
- The model is a decoder-only transformer with an MLP that converts patches of a series into tokens, plus positional encodings.
- Output patches can be longer than input patches.
- Training cost reported: TPUv5e with 16 tensor cores for ~2 days for the 200M-parameter model; one estimate equates this to ~60 GPU-hours on 8×A100, seen as modest compared to LLMs.
Universality vs. domain specificity
- Some find a general time-series model conceptually odd: how can one model handle egg prices, inflation, stocks, etc.?
- Others argue it doesn’t “understand” domains but learns generic structures: trend, seasonality, residuals, and cross-domain patterns linked to human behavior, weather, holidays.
- Synthetic training data based on simple statistical models (piecewise linear, ARMA, sine/cosine seasonality) is cited as a way to encode universal temporal patterns.
- Comparisons are made to LLMs and to generic compressors like JPEG: same machinery, many content types.
Practical performance vs. traditional methods
- One reported internal test finds TimesFM performs about as well as ARIMA on their data but is heavier and slower, making its niche unclear when a data scientist can just fit ARIMA/related models.
- Several note that in time-series competitions, traditional methods (ARIMA, LightGBM, etc.) often match or beat deep nets, except in specific setups.
- A linked critical essay argues against time-series foundation models; some investors are portrayed as perhaps over-optimistic.
Use cases, limits, and skepticism
- Suggested good targets: relatively predictable series (insurance mortality, electricity demand, advertising campaign performance).
- Strong skepticism about using such models for chaotic domains like Bitcoin or “breaking” stock markets.
- Debate over whether “universal” forecasting is meaningful given chaos, limited information, and feedback effects from widespread forecasting itself.
- Some propose alternative workflows, e.g., using an LLM plus classical stats tools to automatically design traditional forecasting models.