To address the challenges of deploying large-scale language models (LLMs) in resource-constrained environments, this paper proposes NoWag (Normalized Weight and Activation Guided Compression), a unified one-shot shape-preserving compression algorithm framework. NoWag compresses Llama-2 (7B, 13B, 70B) and Llama-3 (8B, 70B) models using two shape-preserving techniques: vector quantization (NoWag-VQ) and unstructured/semi-structured pruning (NoWag-P). Experimental results demonstrate that NoWag-VQ significantly outperforms state-of-the-art one-shot vector quantization methods, and NoWag-P is competitive with leading pruning techniques. This highlights the commonalities between the two compression paradigms and suggests promising directions for future research. The source code is available on GitHub.