This paper presents hdl2v, a novel dataset for improving the performance of large-scale language models (LLMs) for Verilog code generation. hdl2v aims to augment the amount of publicly available Verilog data by translating three hardware description languages, VHDL, Chisel, and PyMTL3, to Verilog. We demonstrate that the hdl2v dataset improves the VerilogEvalV2 performance of a 32 billion-parameter open-weight model by up to 23% (pass@10), and that a data augmentation-based fine-tuning approach also improves the performance by up to 63%. We also analyze and present the characteristics of the HDL-to-Verilog dataset for future work.