Hongyu Li

h-index: 10 693 citations 26 papers (total)

Papers in Database (2)

defense arXiv Jan 19, 2026 · 11w ago

LSSF: Safety Alignment for Large Language Models through Low-Rank Safety Subspace Fusion

Guanghao Zhou, Panjia Qiu, Cen Chen et al. · East China Normal University · Ant Group

Post-hoc LLM safety re-alignment via low-rank safety subspace fusion to restore guardrails degraded by fine-tuning

Transfer Learning Attack Prompt Injection nlp
3 citations 1 influentialPDF
attack arXiv Nov 23, 2025 · Nov 2025

Shadows in the Code: Exploring the Risks and Defenses of LLM-based Multi-Agent Software Development Systems

Xiaoqing Wang, Keman Huang, Bin Liang et al. · Renmin University of China · Ant Group

Attacks LLM multi-agent software dev systems via prompt injection and compromised agents to produce hidden malware

Prompt Injection Excessive Agency nlp
PDF Code