Junfeng Fang

Papers in Database (2)

defense arXiv Feb 27, 2026 · 5w ago

GuardAlign: Test-time Safety Alignment in Multimodal Large Language Models

Xingyu Zhu, Beier Zhu, Junfeng Fang et al. · University of Science and Technology of China · Nanyang Technological University +2 more

Training-free defense for VLMs uses optimal transport patch detection and attention calibration to block visual jailbreaks

Input Manipulation Attack Prompt Injection visionnlpmultimodal
PDF
defense arXiv Feb 12, 2026 · 7w ago

SafeNeuron: Neuron-Level Safety Alignment for Large Language Models

Zhaoxin Wang, Jiaming Liang, Fengbin Zhu et al. · Xidian University · National University of Singapore +1 more

Defends LLM safety alignment against neuron pruning attacks by redistributing safety representations across the network via selective neuron freezing

Prompt Injection nlpmultimodal
PDF