MEETI (MIMIC-IV-Ext ECG-Text-Image) is the first large-scale multimodal ECG dataset that synchronizes electrocardiogram (ECG) waveform data, high-resolution ECG images, and detailed interpretation text generated by a large-scale language model. Each MEETI record consists of four components: raw ECG waveform, corresponding plot image, extracted feature parameters, and detailed interpretation text, which are aligned using a consistent unique identifier. This integrated structure supports transformer-based multimodal learning and enables fine-grained and interpretable inference on cardiac health. By bridging the gap between traditional signal analysis, image-based interpretation, and language-based understanding, MEETI lays a solid foundation for explainable next-generation multimodal cardiovascular AI and provides the research community with a comprehensive benchmark for the development and evaluation of ECG-based AI systems.