This paper compares the performance of imitation learning (IL) and reinforcement learning (RL) for surgical action planning, which predicts future surgical actions (instrument-verb-target triplet) in laparoscopic surgery. Using the CholecT50 dataset, we compared and evaluated imitation learning-based Dual-task Autoregressive Imitation Learning (DARIL) with three reinforcement learning variants (world-model-based RL, direct video RL, and inverse reinforcement learning-enhanced). Results show that all reinforcement learning techniques underperform imitation learning-based DARIL (e.g., world-model RL achieved 3.1% mAP after 10 seconds), and distribution matching on the expert-annotated test set tends to favor imitation learning. This finding challenges the conventional assumption of the superiority of reinforcement learning in sequential decision-making.