ICML 2023 – A Watermark for Large Language Models


In this episode we discuss A Watermark for Large Language Models
by John Kirchenbauer, Jonas Geiping, Yuxin Wen, Jonathan Katz, Ian Miers, Tom Goldstein. This paper presents a watermarking framework for large language models (LLMs), aiming to embed hidden signals in the generated text while remaining undetectable to humans. The approach involves selecting specific tokens and promoting their use in text generation. The proposed watermark is tested on a multi-billion-parameter LLM and its robustness and security are discussed, highlighting the need to detect and audit machine-generated text to mitigate potential malicious use.


Posted

in

by

Tags: