Watermarking Large Language Models to Fight Plagiarism with Tom Goldstein
EPISODE 621
|
MARCH
20,
2023
Watch
Follow
Share
About this Episode
Today we’re joined by Tom Goldstein, an associate professor at the University of Maryland. Tom’s research sits at the intersection of ML and optimization and has previously been featured in the New Yorker for his work on invisibility cloaks, clothing that can evade object detection. In our conversation, we focus on his more recent research on watermarking LLM output. We explore the motivations behind adding these watermarks, how they work, and different ways a watermark could be deployed, as well as political and economic incentive structures around the adoption of watermarking and future directions for that line of work. We also discuss Tom’s research into data leakage, particularly in stable diffusion models, work that is analogous to recent guest Nicholas Carlini’s research into LLM data extraction.
About the Guest
Tom Goldstein
University of Maryland
Resources
- A Watermark for Large Language Models - Twitter Thread
- Dressing for the Surveillance Age | The New Yorker
- Adversarial Attacks on Machine Learning Systems for High-Frequency Trading
- Adversarial attacks on Copyright Detection Systems
- Diffusion Art or Digital Forgery? Investigating Data Replication in Diffusion Models
- Cold Diffusion: Inverting Arbitrary Image Transforms Without Noise
