Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

UltraSTF: Ultra-Compact Model for Large-Scale Spatio-Temporal Forecasting

Created by
  • Haebom

Author

Chin-Chia Michael Yeh, Xiran Fan, Zhimeng Jiang, Yujie Fan, Huiyuan Chen, Uday Singh Saini, Vivian Lai, Xin Dai, Junpeng Wang, Zhongfang Zhuang, Liang Wang, Yan Zheng

Outline

This paper proposes a novel prediction model, UltraSTF, to address the high dimensionality of spatiotemporal data. Existing SparseTSF models leverage periodicity to reduce model size, but suffer from a limitation in properly capturing temporal dependencies within periods. UltraSTF maintains the advantages of SparseTSF while incorporating an ultra-compact shape bank component to effectively learn intra-cycle dynamics. This utilizes an attention mechanism to efficiently capture recurring patterns in temporal time series. As a result, UltraSTF achieves state-of-the-art performance on the LargeST benchmark, while extending the Pareto frontier of existing approaches by using less than 0.2% of the parameters compared to the second-best model.

Takeaways, Limitations

Takeaways:
Introducing UltraSTF, a new state-of-the-art model for spatiotemporal data prediction.
Effectively solves the Limitations (insufficient temporal dependence within a cycle) issue of the existing model SparseTSF.
Achieving high predictive performance with very few parameters (Pareto frontier extension)
Efficient intra-cycle pattern learning using attention mechanisms
Limitations:
Performance verification on datasets other than the LargeST benchmark is needed.
Further research is needed on the model's complexity and interpretability.
Possible lack of detailed description of the design and optimization of micro-format banking components.
👍