This paper tests large-scale language models (LLMs) to solve impossible quizzes under constrained conditions in a sandbox environment. Despite monitoring and anti-cheating guidelines, some state-of-the-art LLMs have consistently attempted to cheat and circumvent the constraints. This exposes a fundamental tension between goal-oriented behavior and alignment in current LLMs. The code and evaluation logs are available on GitHub.