This paper presents an efficient, real-time, and deployable deep learning-based direction prediction method using only visual input from a mobile device to address indoor navigation challenges where GPS access is challenging. Specifically, we automate, enhance, and robust the data collection, annotation, and training processes through a novel graph-based path generation method, explainable data augmentation, and curriculum learning. Using a large-scale indoor shopping mall video dataset, we annotate each frame with the correct direction to the target location. Direction prediction is achieved solely using visual information, without the need for specialized sensors, additional path markers, scene map information, or internet access. We will develop and release a user-friendly Android application, along with the data and code.