Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Time2time: Causal Intervention in Hidden States to Simulate Rare Events in Time Series Foundation Models

Created by
  • Haebom

Author

Debdeep Sanyal, Aaryan Nagpal, Dhruv Kumar, Murari Mandal, Saurabh Deshpande

Outline

While Transformer-based models excel at predicting everyday patterns, questions remain as to whether they internalize semantic concepts, such as market conditions, or simply fit curves. Furthermore, questions arise about whether these internal representations can be leveraged to simulate rare and risky events, such as market crashes. To address these issues, this paper introduces a causal intervention technique called activation transplantation. This technique manipulates the hidden state by applying statistical moments from one event (e.g., a past crash) to another event (e.g., a period of calm) during a forward pass. This procedure deterministically controls the prediction: injecting crash semantics induces a downward prediction, while injecting calm semantics suppresses the crash and restores stability. Beyond binary control, we find that the model encodes a notion of event severity, and that the latent vector norm is directly correlated with the magnitude of the system shock. Validated on two architectures (Toto and Chronos), the technique demonstrates that manipulable and semantically informed representations are powerful properties for large-scale time series transformers.

Takeaways, Limitations

Takeaways:
We demonstrate that the model can internalize semantic concepts such as market conditions and manipulate them to control predictions.
Enables "what-if" analysis through activation transplant techniques, which can be used for strategic stress testing.
Transforming the interpretability of model internal representations from post hoc attribution to direct causal intervention.
The robustness of the results is ensured by validation on two different architectures (Toto, Chronos).
Limitations:
The paper does not explicitly present specific Limitations information (but does provide general Limitations: model complexity, data dependence, generalizability to specific markets, etc.)
👍