Latest papers

7 papers
attack arXiv Dec 18, 2025 · Dec 2025

In-Context Probing for Membership Inference in Fine-Tuned Language Models

Zhexi Lu, Hongliang Chi, Nathalie Baracaldo et al. · Rensselaer Polytechnic Institute · IBM Research +1 more

Attacks fine-tuned LLM privacy via in-context probing to infer training membership without shadow model training

Membership Inference Attack nlp
PDF
attack arXiv Dec 14, 2025 · Dec 2025

COBRA: Catastrophic Bit-flip Reliability Analysis of State-Space Models

Sanjay Das, Swastik Bhattacharya, Shamik Kundu et al. · University of Texas at Dallas · Intel Corporation +1 more

Bit-flip attack framework shows flipping one bit in Mamba-1.4b collapses accuracy from 74.64% to 0% on LAMBADA

Model Poisoning nlp
PDF Code
defense arXiv Nov 11, 2025 · Nov 2025

WaterMod: Modular Token-Rank Partitioning for Probability-Balanced LLM Watermarking

Shinwoo Park, Hyejin Park, Hyeseon Ahn et al. · Yonsei University · Rensselaer Polytechnic Institute

Watermarks LLM text outputs via modular token-rank partitioning, supporting binary and multi-bit provenance tracing without fluency loss

Output Integrity Attack nlp
4 citations PDF Code
attack arXiv Oct 25, 2025 · Oct 2025

Power to the Clients: Federated Learning in a Dictatorship Setting

Mohammadsajad Alipour, Mohammad Mohammadi Amiri · Rensselaer Polytechnic Institute

Introduces 'dictator clients' in federated learning that erase all benign participant contributions while preserving their own through crafted model updates

Data Poisoning Attack visionnlpfederated-learning
PDF
defense arXiv Oct 10, 2025 · Oct 2025

A Linguistics-Aware LLM Watermarking via Syntactic Predictability

Shinwoo Park, Hyejin Park, Hyeseon Ahn et al. · Yonsei University · Rensselaer Polytechnic Institute

Linguistics-aware LLM text watermarking using POS n-gram entropy to balance quality and detectability without model logit access

Output Integrity Attack nlp
PDF Code
defense arXiv Sep 11, 2025 · Sep 2025

CryptGNN: Enabling Secure Inference for Graph Neural Networks

Pritam Sen, Yao Ma, Cristian Borcea · New Jersey Institute of Technology · Rensselaer Polytechnic Institute

SMPC-based secure GNN inference framework that protects model parameters from clients and client inputs from cloud providers

Model Theft graph
PDF
defense arXiv Aug 6, 2025 · Aug 2025

SenseCrypt: Sensitivity-guided Selective Homomorphic Encryption for Joint Federated Learning in Cross-Device Scenarios

Borui Li, Li Yan, Junhao Han et al. · Xi’an Jiaotong University · Rensselaer Polytechnic Institute

Selective homomorphic encryption for federated learning that defends against gradient inversion attacks while cutting training overhead by up to 88.7%

Model Inversion Attack federated-learning
PDF