We study the phenomenon of evaluation awareness for the Llama-3.3-70B-Instruct model. Evaluation awareness refers to the ability of a language model to distinguish between a test phase and a deployment phase, and has serious safety and policy implications that could undermine the trustworthiness of AI governance frameworks and voluntary industry commitments. In this paper, we show that linear probes can be used to distinguish between true evaluation prompts and deployment prompts, suggesting that the current model internally represents this distinction. We also find that current safety assessments are accurately classified by the probes, suggesting that they already appear artificial or untrue to the model. These results highlight the importance of ensuring trustworthy assessments and understanding deceptive features. More broadly, our study demonstrates how model internals can be leveraged to support black-box safety audits, especially for future models that are more adept at evaluation awareness and deception.