This paper studies the phenomenon of language switching in a Chinese-English bilingual inference model. The results reveal that language switching is a strategic behavior that enhances inference ability. Specifically, the reinforcement learning-based training phase induces language mixing, and suppressing language mixing reduces accuracy. Furthermore, we demonstrate that accuracy can be improved by guiding decoding using lightweight probes that predict the benefit of language switching.