This paper presents VN-MTEB, a large-scale benchmark dataset for evaluating Vietnamese embedding models. Vietnam's high internet usage and prevalence of online toxicity make embedding models crucial. However, to address the lack of a suitable evaluation dataset, we translated the existing English Massive Text Embedding Benchmark (MTEB) into Vietnamese. Leveraging large-scale language models (LLMs) and state-of-the-art embedding models, we achieved high-quality translation and filtering, preserving natural language flow and semantic accuracy, even preserving Named Entity Recognition (NER) and code fragments. Finally, we present VN-MTEB, a dataset comprised of 41 datasets across six tasks. Analysis results show that large, complex models using Rotary Positional Embeddings outperform models using Absolute Positional Embeddings. The dataset is publicly available on HuggingFace.