Latest papers

3 papers
benchmark arXiv Feb 13, 2026 · 7w ago

A Calibrated Memorization Index (MI) for Detecting Training Data Leakage in Generative MRI Models

Yash Deo, Yan Jia, Toni Lassila et al. · University of York · University of Leeds +3 more

Proposes calibrated memorization metrics using MRI foundation model features to detect training data duplication in generative MRI models

Model Inversion Attack vision
PDF Code
benchmark arXiv Oct 12, 2025 · Oct 2025

Safeguarding Efficacy in Large Language Models: Evaluating Resistance to Human-Written and Algorithmic Adversarial Prompts

Tiarnaigh Downey-Webb, Olamide Jogunola, Oluwaseun Ajao · Manchester Metropolitan University

Benchmarks four LLMs against gradient-based and prompt-level jailbreaks, revealing cross-model transferability of GCG and TAP attacks

Input Manipulation Attack Prompt Injection nlp
PDF
benchmark arXiv Sep 5, 2025 · Sep 2025

Differential Robustness in Transformer Language Models: Empirical Evaluation Under Adversarial Text Attacks

Taniya Gidatkar, Oluwaseun Ajao, Matthew Shardlow · Manchester Metropolitan University

Benchmarks BERT, RoBERTa, and Flan-T5 robustness against word-substitution adversarial attacks, revealing stark vulnerability differences

Input Manipulation Attack nlp
PDF