defense 2026

Online LLM watermark detection via e-processes

Weijie Su 1, Ruodu Wang 2, Zinan Zhao 3

0 citations · 32 references · arXiv (Cornell University)

α

Published on arXiv

2602.14286

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

Proposed e-process framework achieves competitive or superior watermark detection power compared to existing fixed-sample methods while additionally providing anytime-valid sequential guarantees.

e-process watermark detection

Novel technique introduced


Watermarking for large language models (LLMs) has emerged as an effective tool for distinguishing AI-generated text from human-written content. Statistically, watermark schemes induce dependence between generated tokens and a pseudo-random sequence, reducing watermark detection to a hypothesis testing problem on independence. We develop a unified framework for LLM watermark detection based on e-processes, providing anytime-valid guarantees for online testing. We propose various methods to construct empirically adaptive e-processes that can enhance the detection power. In addition, theoretical results are established to characterize the power properties of the proposed procedures. Some experiments demonstrate that the proposed framework achieves competitive performance compared to existing watermark detection methods.


Key Contributions

  • Unified e-process framework for LLM watermark detection with anytime-valid Type I error control under arbitrary stopping times, enabling sequential/streaming detection
  • Empirically adaptive e-process construction methods (adaptive weights, online Grenander algorithm for calibrators) that enhance detection power, sometimes outperforming non-sequential baselines
  • Asymptotic power-one theoretical results characterizing the power properties of the proposed procedures under the Gumbel-max watermark scheme

🛡️ Threat Analysis

Output Integrity Attack

Directly addresses detection of watermarks embedded in LLM-generated text outputs to distinguish AI-generated from human-written content — a content provenance and output integrity problem. The paper's sole contribution is the watermark detection framework, not model ownership or adversarial input manipulation.


Details

Domains
nlp
Model Types
llm
Threat Tags
inference_time
Applications
ai-generated text detectionllm watermark verificationstreaming text authentication