This paper presents the Semantic Occupancy Prediction (SOP) task, which simultaneously infers scene geometry and semantic labels from images to address the challenges of object occlusion and incomplete scene data in autonomous driving perception systems. To address the shortcomings of existing camera-based methods, which treat all categories equally and rely primarily on local features, resulting in poor prediction performance, especially for dynamic foreground objects, we propose an Object-Centric SOP (OC-SOP) framework. OC-SOP significantly improves foreground object prediction accuracy by incorporating high-level object-centric cues extracted through detection branches into the semantic occupancy prediction pipeline, achieving state-of-the-art performance across all categories on the SemanticKITTI dataset.