This paper studies how effectively a Large-Scale Language Model (LLM) leverages the context of diverse languages in Augmented Retrieval Generation (RAG) systems. Specifically, we evaluate the LLM's ability to leverage relevant phrases in languages other than the query, respond in the expected language, and focus on relevant phrases even when presented with "interrupting" phrases in multiple languages. Experiments using four LLMs and three QA datasets spanning 48 languages reveal that while LLM excels at extracting information from phrases in languages other than the query, it struggles to generate complete answers in the correct language. Furthermore, we find that interrupting phrases negatively impact answer quality regardless of language, with interrupting phrases in the query language having a greater impact.