This paper proposes GVote, an adaptive KV-cache compression technique, to address the growing memory footprint of KV-caches, which are used to accelerate autoregressive decoding of large-scale language models (LLMs). Unlike existing methods that use a fixed compression ratio, GVote dynamically determines the optimal cache size by predicting the attention demand of future queries through Monte-Carlo sampling. Experiments on various benchmarks, including GSM8K, RULER, and Longbench, demonstrate that GVote reduces memory usage by a factor of two while maintaining comparable or higher accuracy compared to existing methods.