This paper aims to improve the performance of large-scale language models (LLMs) for Verilog code generation by presenting the hdl2v dataset, which converts hardware description languages such as VHDL, Chisel, and PyMTL3 to Verilog. Using the hdl2v dataset, we improve the VerilogEvalV2 performance of a 32 billion parameter open weight model by up to 23% (pass@10), and also improve the performance of a data augmentation-based fine-tuning approach by 63%. We also provide feature analysis of the HDL-to-Verilog dataset for future performance improvements.