Paper 2024/1424
A Waterlog for Detecting and Tracing Synthetic Text from Large Language Models
Abstract
We propose waterlogs, a new direction to detect and trace synthetic text outputs from large language models based on transparency logs. Waterlogs offer major categorical advantages over watermarking: it (1) allows for the inclusion of arbitrary metadata to facilitate tracing, (2) is publicly verifiable by third parties, and (3) operates in a distributed manner while remaining robust and efficient. Waterlogs rely on a verifiable Hamming distance index, a novel data structure that we describe, to efficiently search multi-dimensional semantic hashes of natural language embeddings in a verifiable manner. This data structure may be of independent interest. We implement a waterlog, which we call DREDGE, and benchmark it using synthetic text generated by GPT-2 1.5B and OPT-13B; embeddings are generated via OpenAI's text-embedding-ada-002 model. We provide empirical benchmarks on the efficiency of appending text to the log and querying it for matches. We compare our results to watermarking and outline areas for further research.
Metadata
- Available format(s)
- Category
- Applications
- Publication info
- Preprint.
- Keywords
- llmgenerative AIsynthetic texttransparency logswatermarking
- Contact author(s)
-
brimhall @ cs jhu edu
oweller @ cs jhu edu
mgreen @ cs jhu edu
imiers @ umd edu - History
- 2024-09-14: approved
- 2024-09-11: received
- See all versions
- Short URL
- https://ia.cr/2024/1424
- License
-
CC BY
BibTeX
@misc{cryptoeprint:2024/1424, author = {Brennon Brimhall and Orion Weller and Matthew Green and Ian Miers}, title = {A Waterlog for Detecting and Tracing Synthetic Text from Large Language Models}, howpublished = {Cryptology {ePrint} Archive, Paper 2024/1424}, year = {2024}, url = {https://eprint.iacr.org/2024/1424} }