LUMA is a new dataset designed to enhance the reliability of multimodal deep learning, which integrates diverse information sources (text, images, audio, and video) to improve decision making. It extends the existing CIFAR-10/100 dataset by adding audio and text data, and is specifically designed to allow for the infusion of various types and degrees of uncertainty, with the goal of learning from uncertain data. The audio data was extracted from three audio corpora, and the text data was generated using the Gemma-7B LLM. LUMA is available as a Python package that includes functions to generate multiple dataset variations by controlling data diversity, the amount of noise in each modality, and the addition of out-of-distribution samples. It also provides a baseline pretrained model and three uncertainty quantification methods: Monte-Carlo Dropout, Deep Ensemble, and Reliable Conflictive Multi-View Learning. This supports the development, evaluation, and benchmarking of reliable and robust multimodal deep learning methods.