Paper 2024/1424

A Waterlog for Detecting and Tracing Synthetic Text from Large Language Models

Brennon Brimhall, Johns Hopkins University
Orion Weller, Johns Hopkins University
Matthew Green, Johns Hopkins University
Ian Miers, University of Maryland, College Park
Abstract

We propose waterlogs, a new direction to detect and trace synthetic text outputs from large language models based on transparency logs. Waterlogs offer major categorical advantages over watermarking: it (1) allows for the inclusion of arbitrary metadata to facilitate tracing, (2) is publicly verifiable by third parties, and (3) operates in a distributed manner while remaining robust and efficient. Waterlogs rely on a verifiable Hamming distance index, a novel data structure that we describe, to efficiently search multi-dimensional semantic hashes of natural language embeddings in a verifiable manner. This data structure may be of independent interest. We implement a waterlog, which we call DREDGE, and benchmark it using synthetic text generated by GPT-2 1.5B and OPT-13B; embeddings are generated via OpenAI's text-embedding-ada-002 model. We provide empirical benchmarks on the efficiency of appending text to the log and querying it for matches. We compare our results to watermarking and outline areas for further research.

Metadata
Available format(s)
PDF
Category
Applications
Publication info
Preprint.
Keywords
llmgenerative AIsynthetic texttransparency logswatermarking
Contact author(s)
brimhall @ cs jhu edu
oweller @ cs jhu edu
mgreen @ cs jhu edu
imiers @ umd edu
History
2024-09-14: approved
2024-09-11: received
See all versions
Short URL
https://ia.cr/2024/1424
License
Creative Commons Attribution
CC BY

BibTeX

@misc{cryptoeprint:2024/1424,
      author = {Brennon Brimhall and Orion Weller and Matthew Green and Ian Miers},
      title = {A Waterlog for Detecting and Tracing Synthetic Text from Large Language Models},
      howpublished = {Cryptology {ePrint} Archive, Paper 2024/1424},
      year = {2024},
      url = {https://eprint.iacr.org/2024/1424}
}
Note: In order to protect the privacy of readers, eprint.iacr.org does not use cookies or embedded third party content.