Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Screen Hijack: Visual Poisoning of VLM Agents in Mobile Environments

Created by
  • Haebom

Author

Xuan Wang, Siyuan Liang, Zhe Liu, Yi Yu, Yuliang Lu, Xiaochun Cao, Ee-Chien Chang, Xitong Gao

Outline

This paper addresses the security vulnerabilities of Vision-Language Model (VLM)-based mobile agents. In particular, we point out that mobile agents fine-tuned with user-generated datasets are vulnerable to covert backdoor attacks during the training process, and propose a novel clean-label backdoor attack technique called GHOST. GHOST injects malicious behaviors by manipulating only the visual inputs of the training data (without changing the labels or instructions). It is designed to induce attacker-controlled responses when certain visual triggers (static patches, dynamic motion cues, and low-transparency overlays) are recognized, and experimentally demonstrates that it achieves high success rate (up to 94.67%) and high normal task performance (up to 95.85% FSR) on several Android apps and VLM architectures. This study is the first to address the security threats of VLM-based mobile agents and emphasizes the need for effective defense mechanisms in the training pipeline.

Takeaways, Limitations

Takeaways:
First to reveal vulnerability of VLM-based mobile agent to clean label backdoor attack.
Demonstrated the possibility of a backdoor attack with high success rate and concealment through the GHOST attack technique.
Suggests the need for enhanced security for mobile agent training pipelines.
Presents realistic attack scenarios using various visual triggers (static, dynamic, low transparency).
Limitations:
There is little discussion of defense techniques against GHOST attacks.
Although we have evaluated various VLM architectures and apps, further research is needed to determine whether our approach can be generalized to all types of VLMs and apps.
Further validation of generalizability to real-world attack scenarios is needed.
👍