This paper explores how compression techniques such as quantization, distillation, and pruning improve the computational efficiency of large-scale inference models (LRMs). Addressing the limitations of previous research, we compare all three compression techniques and conduct in-depth interpretive analysis. We benchmark the DeepSeek-R1 model on four inference datasets and investigate the impact of compression on inference performance through activation-based, fine-grained causal relationship analysis.