This paper analyzes the phenomenon where low-bit weight-only quantization significantly reduces the memory footprint of large-scale language models (LLMs), but disproportionately impacts certain examples. We analyze LLMs ranging in size from 7 to 70 bits, applying various 3- and 4-bit quantization methods. We find that the quantization errors of 50 pairs of methods exhibit a strong correlation (average 0.82) on the FineWeb example. Furthermore, we demonstrate that the residual stream size of a full-precision model is an indicator of future quantization error. We hypothesize a relationship between residual stream size and error amplification and accumulation across layers. Using LLM localization techniques, early termination, and active patches, we show that examples with large errors rely on precise residual activation in later layers, and that the output of MLP gates plays a crucial role in maintaining perplexity. In conclusion, this study identifies the reasons why large quantization errors occur on certain examples and the most important model components for maintaining performance.