This paper presents PennyLang, a high-quality dataset dedicated to PennyLane, to address the lack of high-quality datasets that limit the utilization of large-scale language models (LLMs) in quantum software development. PennyLang consists of 3,347 PennyLane quantum code samples and contextual descriptions collected from textbooks, official documents, and open-source repositories. This paper presents three contributions: the generation and release of PennyLang, an automated quantum code dataset construction framework, and baseline evaluation using multiple open-source models within the Retrieval-Augmented Generation (RAG) pipeline. Experimental results demonstrate that combining RAG and PennyLang significantly improves the performance of the Qwen 7B and LLaMa 4 models. This contrasts with previous research focused on Qiskit, contributing to the advancement of AI-assisted quantum development by providing LLM-based tools and reproducible methods for PennyLane.