defense arXiv Oct 1, 2025 · Oct 2025
Jiayao Jiang, Bin Liu, Qi Chu et al. · University of Science and Technology of China · Anhui Province Key Laboratory of Digital Security
Novel KAN-based deepfake detector uses facial landmarks to adaptively generate spline activations for artifact detection
Output Integrity Attack vision
The rapid development of deepfake generation techniques necessitates robust face forgery detection algorithms. While methods based on Convolutional Neural Networks (CNNs) and Transformers are effective, there is still room for improvement in modeling the highly complex and non-linear nature of forgery artifacts. To address this issue, we propose a novel detection method based on the Kolmogorov-Arnold Network (KAN). By replacing fixed activation functions with learnable splines, our KAN-based approach is better suited to this challenge. Furthermore, to guide the network's focus towards critical facial areas, we introduce a Landmark-assisted Adaptive Kolmogorov-Arnold Network (LAKAN) module. This module uses facial landmarks as a structural prior to dynamically generate the internal parameters of the KAN, creating an instance-specific signal that steers a general-purpose image encoder towards the most informative facial regions with artifacts. This core innovation creates a powerful combination between geometric priors and the network's learning process. Extensive experiments on multiple public datasets show that our proposed method achieves superior performance.
cnn transformer University of Science and Technology of China · Anhui Province Key Laboratory of Digital Security
defense arXiv Nov 26, 2025 · Nov 2025
Yuxiao Xiang, Junchi Chen, Zhenchao Jin et al. · University of Science and Technology of China · Anhui Province Key Laboratory of Digital Security +1 more
Defends VLMs against unsafe intermediate reasoning by auditing the full Question-Thinking-Answer pipeline with a vision-aware safety guard
Prompt Injection multimodalnlp
Multimodal large reasoning models (MLRMs) are increasingly deployed for vision-language tasks that produce explicit intermediate rationales. However, reasoning traces can contain unsafe content even when the final answer is non-harmful, creating deployment risks. Existing multimodal safety guards primarily evaluate only the input question and the final answer, neglecting the intermediate reasoning process. This oversight allows undetected harm, such as biased inferences or policy-violating use of visual context, to emerge during reasoning. We introduce GuardTrace-VL, a vision-aware safety auditor that monitors the full Question-Thinking-Answer (QTA) pipeline via joint image-text analysis, enabling detection of unsafe content as it emerges in the reasoning stage. To support training and evaluation, we construct the GuardTrace dataset, which is generated through diverse prompting strategies and refined via a MLRM- and human-based voting and verification pipeline. Furthermore, we propose a three-stage progressive training scheme combined with the data refinement process, enabling the model to learn nuanced and context-dependent safety preferences according to different risk levels. On our proposed test set covering both in-domain and out-of-domain scenarios, GuardTrace-VL model achieves an F1 score of 93.1% on unsafe reasoning detection tasks, representing a 13.5% improvement in F1 score compared to the previous strongest multimodal safety defense methods. The codes will be made publicly available.
vlm transformer University of Science and Technology of China · Anhui Province Key Laboratory of Digital Security · The University of Hong Kong