This paper presents a linear probe technique that analyzes the internal activation of AI systems to detect deceptiveness in generated responses. Experiments with the Llama and Qwen models (1.5B to 14B parameters) demonstrated that, especially for larger models with 7B parameters or more, it distinguishes deceptive and non-deceptive responses with an accuracy of over 70-80%. A model fine-tuned with DeepSeek-r1 achieved an accuracy of over 90%. Layer-by-layer analysis revealed a three-stage pattern: detection accuracy was low in the early layers, peaked in the middle layers, and then slightly decreased in later layers. Furthermore, using an iterative null-space projection technique, we identified multiple linear directions that indicate deceptiveness.