This paper presents a method to improve the detailed visual reasoning ability of visual language models (VLMs) even under computationally limited conditions. Inspired by Deepseek-r1, we train small models using Group Relative Policy Optimization (GRPO) and leverage external tools such as zoom. We achieve the greatest benefit by combining GRPO training, a simple reward structure, a streamlined tool call interface, additional token allocation for tool call results, and a mix of training data that overrepresents visually challenging examples. Consequently, we achieve improved performance on some visual question answering (VQA) tasks compared to similarly sized baseline models, thanks to the detailed visual information collected from the external tools.