In this paper, we propose a novel self-supervised learning (SSL) model, EEG-VJEPA, for effective analysis of electroencephalography (EEG) signals. EEG-VJEPA applies the V-JEPA model for video processing to EEG, using joint embedding and adaptive masking to capture spatiotemporal dependencies. Experimental results using the publicly available TUH Abnormal EEG dataset demonstrate that EEG-VJEPA outperforms existing state-of-the-art models in classification accuracy, and can support collaboration with clinicians by capturing physiologically relevant spatiotemporal patterns and providing interpretable embeddings. This demonstrates the potential of EEG-VJEPA as a scalable and reliable EEG analysis framework in real clinical settings.