Less is More: Unlocking Specialization of Time Series Foundation Models via Structured Pruning

Sep 19, 2025·
Lifan Zhao
Lifan Zhao
,
Yanyan Shen
,
Zhaoyang Liu
,
Xue Wang
,
Jiaji Deng
· 0 min read
Abstract
Scaling laws motivate the development of Time Series Foundation Models (TSFMs) that pre-train vast parameters and achieve remarkable zero-shot forecasting performance. Surprisingly, even after fine-tuning, TSFMs cannot consistently outperform smaller, specialized models trained on full-shot downstream data. A key question is how to regularize the adaptation process of TSFMs for a target forecasting task. Through empirical studies on various TSFMs, the pre-trained models often exhibit inherent sparsity and redundancy in computation, suggesting that TSFMs have learned to activate task-relevant network substructures to accommodate diverse forecasting tasks. To preserve this valuable prior knowledge, we propose structured pruning method to regularize the subsequent fine-tuning process by focusing it on a more relevant and compact parameter space. Extensive experiments on seven TSFMs and six benchmarks demonstrate that fine-tuning a smaller, pruned TSFM significantly improves forecasting performance compared to fine-tuning original models. This ``prune-then-finetune’’ paradigm often enables TSFMs to achieve state-of-the-art performance and surpass strong specialized baselines.
Type
Publication
The Thirty-ninth Annual Conference on Neural Information Processing Systems (NeurIPS'25)