LLM Watermarking Using Mixtures and Statistical-to-Computational Gaps
- WaLM
Main:18 Pages
Bibliography:3 Pages
Abstract
Given a text, can we determine whether it was generated by a large language model (LLM) or by a human? A widely studied approach to this problem is watermarking. We propose an undetectable and elementary watermarking scheme in the closed setting. Also, in the harder open setting, where the adversary has access to most of the model, we propose an unremovable watermarking scheme.
View on arXivComments on this paper
