This paper describes "Reinforcement Learning from Augmented Generation (RLAG)," a reinforcement learning-based methodology proposed to address the limited performance of large-scale language models (LLMs) for domain-specific knowledge. RLAG iteratively optimizes the model based on generated outputs, effectively embedding important domain knowledge. It selects the output with the highest log probability and uses three custom reward metrics to guide the optimization process. Experiments on medical, legal, astronomy, and current affairs datasets demonstrate that RLAG outperforms existing methods.