Lihai Nie

Papers in Database (2)

defense arXiv Aug 10, 2025 · Aug 2025

Gradient Surgery for Safe LLM Fine-Tuning

Biao Yi, Jiahao Li, Baolei Zhang et al. · Nankai University

Gradient surgery defense nullifies safety-conflicting gradients during LLM fine-tuning to resist adversarial data poisoning attacks

Data Poisoning Attack Training Data Poisoning nlp
PDF Code
defense arXiv Sep 17, 2025 · Sep 2025

Who Taught the Lie? Responsibility Attribution for Poisoned Knowledge in Retrieval-Augmented Generation

Baolei Zhang, Haoran Xin, Yuxi Chen et al. · Nankai University · University of North Texas +1 more

Detects and attributes poisoned documents in RAG knowledge bases by scoring retrieval ranking, semantics, and generation influence

Data Poisoning Attack Prompt Injection nlp
PDF Code