This paper studies the phenomenon of evaluation awareness in the Llama-3.3-70B-Instruct model. Evaluation awareness refers to the ability of a language model to distinguish between testing and deployment phases, and has serious safety and policy implications that could undermine the reliability of AI governance frameworks and industry-wide voluntary efforts. The researchers demonstrate that linear probes can be used to distinguish between true evaluation and deployment prompts, suggesting that the current model internally represents this distinction. Furthermore, they find that current safety assessments are accurately classified by the probes, suggesting that the model already appears artificial or untrue. These results highlight the importance of ensuring trustworthy assessments and understanding deceptive features. More broadly, this study demonstrates how model internals can be leveraged to support black-box safety audits, especially for future models that are more adept at evaluation awareness and deception.