This paper proposes EvoP, an evolutionary pruning framework, to address the problem of deploying large-scale language models (LLMs) in resource-constrained environments. To address the performance degradation and data feature neglect of existing heuristic-based pruning methods, EvoP introduces a cluster-based calibration dataset sampling (CCDS) strategy to generate diverse calibration datasets and an evolutionary pruning pattern search (EPPS) method to identify optimal pruning patterns. Experiments on various LLMs and subtasks demonstrate the effectiveness of EvoP, demonstrating its practical and scalable solution for deploying LLMs in real-world applications.