Haibo Hu

h-index: 13 576 citations 71 papers (total)

Papers in Database (7)

benchmark WWW Sep 23, 2025 · Sep 2025

MER-Inspector: Assessing model extraction risks from an attack-agnostic perspective

Xinwei Zhang, Haibo Hu, Qingqing Ye et al. · Hong Kong Polytechnic University · Ltd.

Proposes NTK-based theoretical metrics to quantify model extraction risk across architectures without assuming a specific attack strategy

Model Theft vision
4 citations PDF
defense arXiv Nov 29, 2025 · Nov 2025

Adversarial Signed Graph Learning with Differential Privacy

Haobin Ke, Sen Zhang, Qingqing Ye et al. · The Hong Kong Polytechnic University

Defends signed GNNs against link-stealing attacks using adversarial training and differential privacy with node-level guarantees

Membership Inference Attack graph
PDF Code
attack arXiv Jan 20, 2026 · 10w ago

Diffusion-Guided Backdoor Attacks in Real-World Reinforcement Learning

Tairan Huang, Qingqing Ye, Yulin Jin et al. · The Hong Kong Polytechnic University

Diffusion-generated floor patch triggers bypass real-world safety control stacks to reliably activate backdoors in RL robot policies

Model Poisoning reinforcement-learningvision
PDF
defense arXiv Jan 11, 2026 · 12w ago

United We Defend: Collaborative Membership Inference Defenses in Federated Learning

Li Bai, Junxu Liu, Sen Zhang et al. · The Hong Kong Polytechnic University · PolyU Research Centre for Privacy and Security Technologies in Future Smart Systems

Collaborative FL defense framework that limits local memorization to defeat trajectory-based membership inference attacks

Membership Inference Attack federated-learningvision
PDF Code
attack arXiv Jan 29, 2026 · 9w ago

On the Adversarial Robustness of Large Vision-Language Models under Visual Token Compression

Xinwei Zhang, Hangcheng Liu, Li Bai et al. · The Hong Kong Polytechnic University · Nanyang Technological University +1 more

Proposes CAGE, a compression-aware adversarial attack exposing that token-compressed VLM robustness is systematically overestimated by standard attacks

Input Manipulation Attack visionmultimodal
PDF
attack arXiv Feb 10, 2026 · 7w ago

Understanding and Enhancing Encoder-based Adversarial Transferability against Large Vision-Language Models

Xinwei Zhang, Li Bai, Tianwei Zhang et al. · The Hong Kong Polytechnic University · Nanyang Technological University +1 more

Proposes SGMA, a transferable adversarial visual attack on LVLMs targeting semantically critical regions to disrupt cross-modal grounding

Input Manipulation Attack Prompt Injection visionmultimodalnlp
PDF
defense arXiv Jan 29, 2026 · 9w ago

FIT: Defying Catastrophic Forgetting in Continual LLM Unlearning

Xiaoyu Xu, Minxin Du, Kun Fang et al. · The Hong Kong Polytechnic University · Ant Group

Defends continual LLM unlearning of PII, copyright, and harmful content against adversarial recovery via relearning and quantization attacks

Sensitive Information Disclosure nlp
PDF Code