Seminars and Events
LLM-Powered Predictive Inference with Online Text Time Series
Event Details
Time series predictive inference is an important yet challenging task in economics and business, where existing approaches are often designed for low-frequency, survey-based data. With the recent advances of large language models (LLMs), there is growing potential to leverage high-frequency online text data for improved time series prediction, an area still largely unexplored. This paper proposes the LLM-TS, an LLM-based approach for time series predictive inference incorporating online text data. The LLM-TS is based on a joint time series framework that combines survey-based low-frequency data with LLM-generated high-frequency surrogates. The framework relies only on an error correlation assumption, combining a text-embedding-augmented ARX model for the observed gold-standard measurements with a VARX model for the LLM-generated surrogates. LLM-TS employs LLMs such as ChatGPT and the trained BERT models to construct LLM surrogates. Online text embeddings are extracted via LDA and BERT. We establish the asymptotic properties of the method and provide two forms of constructed prediction intervals. We also extend LLM-TS to incorporate deep learning backbones. To demonstrate the practical power of LLM-TS, we apply it to a critical real-world application: inflation forecast. We construct two large high-frequency online text data sets from the U.S. and China, and use LLMs to extract inflation-related signals from texts that reflect price dynamics. The finite-sample performance and practical advantages of LLM-TS are illustrated through extensive simulations and two noisy real data examples, highlighting its potential to improve time series prediction in economic applications. This is a joint work with Yingying Fan, Ao Sun and Yurou Wang.
Zoom connection
Meeting ID: 927 5174 4356
Passcode: 2025
Host: DJ Ashok
POC: Maura Covaci