Yinan Zhong (Zhejiang University), Qianhao Miao (Zhejiang University), Yanjiao Chen (Zhejiang University), Jiangyi Deng (Zhejiang University), Yushi Cheng (Zhejiang University), Wenyuan Xu (Zhejiang University)

Large Language Models (LLMs) have been integrated into many applications (e.g., web agents) to perform more sophisticated tasks. However, LLM-empowered applications are vulnerable to Indirect Prompt Injection (IPI) attacks, where instructions are injected via untrustworthy external data sources. This paper presents Rennervate, a defense framework to detect and prevent IPI attacks. Rennervate leverages attention features to detect the covert injection at a fine-grained token level, enabling precise sanitization that neutralizes IPI attacks while maintaining LLM functionalities. Specifically, the token-level detector is materialized with a 2-step attentive pooling mechanism, which aggregates attention heads and response tokens for IPI detection and sanitization. Moreover, we establish a fine-grained IPI dataset, FIPI, to be open-sourced to support further research. Extensive experiments verify that Rennervate outperforms 15 commercial and academic IPI defense methods, achieving high precision on 5 LLMs and 6 datasets. We also demonstrate that Rennervate is transferable to unseen attacks and robust against adaptive adversaries.

View More Papers

“These cameras are just like the Eye of Sauron”:...

Shijing He (King’s College London), Yaxiong Lei (University of St Andrews), Xiao Zhan (Universitat Politecnica de Valencia), Ruba Abu-Salma (King’s College London), Jose Such (INGENIO (CSIC-UPV))

Read More

Finding Behavioural Biometrics Scripts on the Web Using Dynamic...

Alexandru Bara (University of Waterloo), Aswad Tariq (University of Waterloo), Urs Hengartner (University of Waterloo)

Read More

On Borrowed Time: Measurement-Informed Understanding of the NTP Pool's...

Robert Beverly (San Diego State University), Erik Rye (Johns Hopkins University)

Read More