Kronos: A Foundation Model for the Language of Financial Markets
- URL: http://arxiv.org/abs/2508.02739v1
- Date: Sat, 02 Aug 2025 13:15:59 GMT
- Title: Kronos: A Foundation Model for the Language of Financial Markets
- Authors: Yu Shi, Zongliang Fu, Shuo Chen, Bohan Zhao, Wei Xu, Changshui Zhang, Jian Li,
- Abstract summary: We propose Kronos, a unified, scalable pre-training framework tailored to financial K-line modeling.<n>Kronos discretizes continuous market information into token sequences, preserving both price dynamics and trade activity patterns.<n>We pre-train Kronos using an autoregressive objective on a massive, multi-market corpus of over 12 billion K-line records from 45 global exchanges.
- Score: 42.048823655500115
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The success of large-scale pre-training paradigm, exemplified by Large Language Models (LLMs), has inspired the development of Time Series Foundation Models (TSFMs). However, their application to financial candlestick (K-line) data remains limited, often underperforming non-pre-trained architectures. Moreover, existing TSFMs often overlook crucial downstream tasks such as volatility prediction and synthetic data generation. To address these limitations, we propose Kronos, a unified, scalable pre-training framework tailored to financial K-line modeling. Kronos introduces a specialized tokenizer that discretizes continuous market information into token sequences, preserving both price dynamics and trade activity patterns. We pre-train Kronos using an autoregressive objective on a massive, multi-market corpus of over 12 billion K-line records from 45 global exchanges, enabling it to learn nuanced temporal and cross-asset representations. Kronos excels in a zero-shot setting across a diverse set of financial tasks. On benchmark datasets, Kronos boosts price series forecasting RankIC by 93% over the leading TSFM and 87% over the best non-pre-trained baseline. It also achieves a 9% lower MAE in volatility forecasting and a 22% improvement in generative fidelity for synthetic K-line sequences. These results establish Kronos as a robust, versatile foundation model for end-to-end financial time series analysis. Our pre-trained model is publicly available at https://github.com/shiyu-coder/Kronos.
Related papers
- Deep Learning for Financial Time Series: A Large-Scale Benchmark of Risk-Adjusted Performance [4.889402269887708]
We present a large scale benchmark of modern deep learning architectures for a financial time series prediction and position sizing task.<n>We evaluate linear models, recurrent networks, transformer based architectures, state space models, and recent sequence representation approaches.<n>We find that models explicitly designed to learn rich temporal representations consistently outperform linear benchmarks and generic deep learning models.
arXiv Detail & Related papers (2026-03-02T12:52:50Z) - StockBot 2.0: Vanilla LSTMs Outperform Transformer-based Forecasting for Stock Prices [0.0]
We present an enhanced StockBot architecture that systematically evaluates modern attention-based, convolutional, and recurrent time-series forecasting models.<n>A carefully constructed vanilla LSTM consistently achieves superior predictive accuracy and more stable buy/sell decision-making.
arXiv Detail & Related papers (2026-01-01T04:09:51Z) - Re(Visiting) Time Series Foundation Models in Finance [3.295157175236371]
Financial time series forecasting is central to trading, portfolio optimization, and risk management.<n>Recent advances in time series foundation models (TSFMs) offer a new paradigm for learning generalizable temporal representations from large and diverse datasets.<n>This paper presents the first comprehensive empirical study of TSFMs in global financial markets.
arXiv Detail & Related papers (2025-11-23T18:44:19Z) - Chronos-2: From Univariate to Universal Forecasting [52.753731922908905]
Chronos-2 is a pretrained model capable of handling univariate, multivariate, and covariate-informed forecasting tasks in a zero-shot manner.<n>It delivers state-of-the-art performance across three comprehensive benchmarks: fev-bench, GIFT-Eval, and Chronos Benchmark II.<n>The in-context learning capabilities of Chronos-2 establish it as a general-purpose forecasting model that can be used "as is" in real-world forecasting pipelines.
arXiv Detail & Related papers (2025-10-17T17:00:53Z) - Estimating Time Series Foundation Model Transferability via In-Context Learning [74.65355820906355]
Time series foundation models (TSFMs) offer strong zero-shot forecasting via large-scale pre-training.<n>Fine-tuning remains critical for boosting performance in domains with limited public data.<n>We introduce TimeTic, a transferability estimation framework that recasts model selection as an in-context-learning problem.
arXiv Detail & Related papers (2025-09-28T07:07:13Z) - ByteGen: A Tokenizer-Free Generative Model for Orderbook Events in Byte Space [11.523583937607622]
We introduce ByteGen, a novel generative model that operates directly on the raw byte streams of LOB events.<n>Our work is the complete elimination of feature engineering and tokenization, enabling the model to learn market dynamics from its most fundamental representation.<n>ByteGen successfully reproduces key facts of financial markets, generating realistic price distributions, heavy-tailed returns, and bursty event timing.
arXiv Detail & Related papers (2025-08-04T09:48:42Z) - CTBench: Cryptocurrency Time Series Generation Benchmark [11.576635693346486]
We introduce textsfCTBench, the first comprehensive TSG benchmark tailored for the cryptocurrency domain.<n>textsfCTBench curates an open-source dataset from 452 tokens and evaluates TSG models across 13 metrics spanning 5 key dimensions.<n>We benchmark eight representative models from five methodological families over four distinct market regimes, uncovering trade-offs between statistical fidelity and real-world profitability.
arXiv Detail & Related papers (2025-08-03T17:07:08Z) - Time Series Foundation Models for Multivariate Financial Time Series Forecasting [0.0]
Time Series Foundation Models (TSFMs) offer a promising solution through pretraining on diverse time series corpora.<n>This study evaluates two TSFMs across three financial forecasting tasks: US 10-year Treasury yield changes, EUR/USD volatility, and equity spread prediction.
arXiv Detail & Related papers (2025-07-09T21:43:06Z) - DELPHYNE: A Pre-Trained Model for General and Financial Time Series [2.601248228220401]
Time-series data is valuable in financial applications, where it helps in detecting patterns, understanding market behavior, and making informed decisions based on historical data.<n>Recent advances in language modeling have led to the rise of time-series pre-trained models that are trained on vast collections of datasets and applied to diverse tasks across financial domains.<n>However, existing time-series pre-trained models have not shown boosts in performance over simple finance benchmarks in both zero-shot and fine-tuning settings.
arXiv Detail & Related papers (2025-05-12T16:53:29Z) - FinTSB: A Comprehensive and Practical Benchmark for Financial Time Series Forecasting [58.70072722290475]
Financial time series (FinTS) record the behavior of human-brain-augmented decision-making.<n>FinTSB is a comprehensive and practical benchmark for financial time series forecasting.
arXiv Detail & Related papers (2025-02-26T05:19:16Z) - Sundial: A Family of Highly Capable Time Series Foundation Models [64.6322079384575]
We introduce Sundial, a family of native, flexible, and scalable time series foundation models.<n>Our models are pre-trained without specifying any prior distribution and can generate multiple probable predictions.<n>Sundial achieves state-of-the-art results on both point and probabilistic forecasting benchmarks with a just-in-time inference speed.
arXiv Detail & Related papers (2025-02-02T14:52:50Z) - GIFT-Eval: A Benchmark For General Time Series Forecasting Model Evaluation [90.53485251837235]
Time series foundation models excel in zero-shot forecasting, handling diverse tasks without explicit training.
GIFT-Eval is a pioneering benchmark aimed at promoting evaluation across diverse datasets.
GIFT-Eval encompasses 23 datasets over 144,000 time series and 177 million data points.
arXiv Detail & Related papers (2024-10-14T11:29:38Z) - PLUTUS: A Well Pre-trained Large Unified Transformer can Unveil Financial Time Series Regularities [0.848210898747543]
Financial time series modeling is crucial for understanding and predicting market behaviors.
Traditional models struggle to capture complex patterns due to non-linearity, non-stationarity, and high noise levels.
Inspired by the success of large language models in NLP, we introduce $textbfPLUTUS$, a $textbfP$re-trained $textbfL$arge.
PLUTUS is the first open-source, large-scale, pre-trained financial time series model with over one billion parameters.
arXiv Detail & Related papers (2024-08-19T15:59:46Z) - Chronos: Learning the Language of Time Series [79.38691251254173]
Chronos is a framework for pretrained probabilistic time series models.
We show that Chronos models can leverage time series data from diverse domains to improve zero-shot accuracy on unseen forecasting tasks.
arXiv Detail & Related papers (2024-03-12T16:53:54Z) - Low-Rank Temporal Attention-Augmented Bilinear Network for financial
time-series forecasting [93.73198973454944]
Deep learning models have led to significant performance improvements in many problems coming from different domains, including prediction problems of financial time-series data.
The Temporal Attention-Augmented Bilinear network was recently proposed as an efficient and high-performing model for Limit Order Book time-series forecasting.
In this paper, we propose a low-rank tensor approximation of the model to further reduce the number of trainable parameters and increase its speed.
arXiv Detail & Related papers (2021-07-05T10:15:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.