This paper introduces AFABench, the first benchmark framework for systematically evaluating Active Feature Acquisition (AFA) methods. AFA selectively acquires features from a subset of data instances, attempting to trade off predictive performance and acquisition cost. AFABench encompasses a variety of synthetic and real-world datasets, supports a variety of acquisition policies, and offers a modular design for easy integration of new methods and tasks. We implement and evaluate representative algorithms, including static, greedy, and reinforcement learning-based approaches, and introduce AFAContext, a novel synthetic dataset designed to expose the limitations of greedy selection. The results reveal key trade-offs between various AFA strategies and provide actionable insights for future research. The benchmark code is available on GitHub.