To address the challenges of assessing the quality of AI-generated clinical notes, this paper proposes a pipeline that systematically extracts real-world user feedback into a structured checklist. This checklist is designed to be interpretable, based on human feedback, and applicable to LLM-based evaluators. Experiments using over 21,000 clinical records demonstrate that the proposed checklist outperforms existing evaluation methods.