This paper explores the necessity of a world model for flexible, goal-oriented behavior. We demonstrate that any agent capable of generalizing to multi-stage, goal-oriented tasks must learn a predictive model of the environment. This model can be derived from the agent's policy, and we demonstrate that improved agent performance or increasing complexity of achievable goals necessitates the learning of increasingly accurate world models.