This paper proposes SlotMatch, a knowledge distillation framework that effectively transfers object-centric representations to a lightweight student model for unsupervised video segmentation. To overcome the limitations of existing slot attention-based models, which are computationally expensive, SlotMatch aligns teacher and student slots using cosine similarity and operates without additional distillation objectives or auxiliary supervision. Theoretical and experimental evidence demonstrate the unnecessary integration of additional loss functions. Experimental results demonstrate that the SlotMatch-based student model performs equally or better than the best-performing teacher model, SlotContrast, while requiring 3.6x fewer parameters and being 1.9x faster. Furthermore, it outperforms existing unsupervised video segmentation models.