This study examines the trade-off between translation quality and efficiency of large-scale language models (LLMs), using machine translation as a case study. We compared the performance of full, distilled, and quantized models using the Flores+ benchmark and human evaluations on conversational translation tasks in French, Hindi, and Kannada. The 3.3B FP32 model achieved the highest BLEU score but incurred the largest environmental footprint. The distilled 600M FP32 model reduced inference time by 71-78% and carbon emissions by 63-65%, while minimizing the decrease in BLEU score. Aggressive quantization (INT4) also maintained high levels of accuracy and fluency.