Latest papers

3 papers
benchmark arXiv Feb 6, 2026 · 8w ago

Malicious Agent Skills in the Wild: A Large-Scale Security Empirical Study

Yi Liu, Zhihao Chen, Yanjun Zhang et al. · Quantstamp · Fujian Normal University +4 more

Empirical study of 98,380 LLM agent skills finds 157 malicious ones using supply chain theft and instruction hijacking

AI Supply Chain Attacks Insecure Plugin Design Prompt Injection nlp
2 citations 1 influentialPDF
attack arXiv Dec 29, 2025 · Dec 2025

EquaCode: A Multi-Strategy Jailbreak Approach for Large Language Models via Equation Solving and Code Completion

Zhen Liang, Hai Huang, Zhengkui Chen · Zhejiang Sci-Tech University

Jailbreaks LLMs by encoding harmful intent as math equations requiring code-based solutions, achieving 91% success on GPT in one query

Prompt Injection nlp
1 citations PDF Code
tool arXiv Nov 17, 2025 · Nov 2025

Unlocking the Forgery Detection Potential of Vanilla MLLMs: A Novel Training-Free Pipeline

Rui Zuo, Qinyue Tong, Zhe-Ming Lu et al. · Zhejiang University · Zhejiang Sci-Tech University

Training-free MLLM pipeline detects and localizes image forgeries including deepfakes and AIGC edits without additional training

Output Integrity Attack visionmultimodal
PDF