This paper presents Contrastive Reasoning Prompt Optimization (CRPO), a novel framework for prompt optimization in large-scale language models (LLMs). CRPO utilizes the inherent reasoning capabilities of LLMs to optimize prompts through a retrieval-augmented reasoning process that learns from contrastive examples. Using the HelpSteer2 dataset, we compare high- and low-quality prompt-response pairs and allow the LLM to improve its prompts through tiered contrastive reasoning and multi-metric contrastive reasoning. Experimental results show that CRPO outperforms existing methods.