Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

SecInfer: Preventing Prompt Injection via Inference-time Scaling

Created by
  • Haebom

Author

Yupei Liu, Yanting Wang, Yuqi Jia, Jinyuan Jia, Neil Zhenqiang Gong

Outline

SecInfer is a prompt injection attack defense system built on \emph{inference time scaling}, a novel paradigm that improves the performance of LLM by allocating additional computational resources during inference. SecInfer consists of two main steps: system-prompt-guided sampling, which generates multiple responses to a given input using various system prompts, and target-task-guided aggregation, which selects the response most likely to perform the intended task. Extensive experiments demonstrate that SecInfer effectively mitigates both conventional and adaptive prompt injection attacks by leveraging additional computation during inference, outperforming state-of-the-art defense systems and existing inference time scaling approaches.

Takeaways, Limitations

Takeaways:
A novel defense method that utilizes computing resources during inference is presented.
Demonstrated effective defense against conventional and adaptive prompt injection attacks.
Outperforms state-of-the-art defense systems and existing inference time scaling approaches.
Limitations:
The specific Limitations is not stated in the abstract. (Please check the original text.)
Additional computing resources are required during inference (potentially increasing computational costs).
👍