The rapid adoption of large language models (LLMs), such as GPT‐4 and Claude 3.5, underscores the need to distinguish LLM‐generated text from human‐written content to mitigate the spread of misinformation and misuse in education. One promising approach to address this issue is the watermark technique, which embeds subtle statistical signals into LLM‐generated text to enable reliable identification. In this paper, we first generalize the likelihood‐based LLM detection method by introducing a flexible weighted formulation and further adapt this approach to the inverse transform sampling method. Moving beyond watermark detection, we extend this adaptive detection strategy to tackle the more challenging problem of segmenting a given text into watermarked and non‐watermarked substrings. In contrast to the approach in a previous study, which relies on accurate estimation of next‐token probabilities that are highly sensitive to prompt estimation, our proposed framework removes the need for precise prompt estimation. Extensive numerical experiments demonstrate that the proposed methodology is both effective and robust in accurately segmenting texts containing a mixture of watermarked and non‐watermarked content.
Adaptive Testing for Segmenting Watermarked Texts From Language Models
Xingchi Li,Xiaochi Liu,Guanxun Li
Published 2025 in Stat
ABSTRACT
PUBLICATION RECORD
- Publication year
2025
- Venue
Stat
- Publication date
2025-11-10
- Fields of study
Mathematics, Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-44 of 44 references · Page 1 of 1
CITED BY
Showing 1-1 of 1 citing papers · Page 1 of 1