In this paper, we present the first systematic characterization of the KV$ workload patterns of real-world LLM service providers, given that intermediate result caching (KV$) plays a critical role in improving performance in large-scale language model (LLM) service provisioning, but system design decisions such as cache eviction policies are highly workload-dependent. We make several observations that have not been addressed in previous studies focusing on synthetic workloads: that KV$ reuse across requests is asymmetric, that reuse across single-turn requests is as important as reuse across multi-turn requests, that reuse times and probabilities vary across all requests, but patterns are predictable for specific request categories, and that the overall cache size is appropriate for an ideal cache hit ratio. Based on these characterizations, we propose a workload-aware cache eviction policy that improves service performance on real-world trace data, especially when cache capacity is limited.