Watermark Anything
Perceived purpose and significance
- Many see this as an invisible watermarking system aimed at generative AI images.
- Claimed robustness to common transformations (cropping, rotation, brightness changes, recompression, inpainting, splicing) is viewed as the main technical advance.
- Potential uses mentioned: provenance of AI content, copyright tracing, camera-origin signatures, and tracking leaks of proprietary media.
Robustness and attack surface
- Past invisible watermarks were often broken by simple operations like resize, slight rotation, or re-watermarking (“double watermark” attacks).
- If this system resists those, commenters consider it a noteworthy improvement.
- There is skepticism that any method is unbreakable; references to existing “watermark attacker” tools and trivial “destroy it with noise” strategies.
- Practical friction: some users report difficulty running the repo due to ML tooling (CUDA/PyTorch) issues, with suggestions to use Colab.
Use in AI training and provenance
- One proposed driver: help AI providers filter out AI-generated images from future training data to avoid “model collapse” from self-training.
- Others speculate about watermark schemes evolving over time, potential for latent watermark patterns to be learned by downstream models, and ideas like Merkle-tree-like registries.
- A suggested strategy is to exclude watermarked inputs from training, which could conflict with upcoming cryptographically signed camera images.
Potential for misuse and governance concerns
- Strong concern that robust, ubiquitous watermarking aids repressive or merely self-interested governments/corporations in deanonymizing leakers, whistleblowers, and dissidents.
- Historical analogies: printer tracking dots and attempts to trace message forwarding chains.
- Debate over whether even a “virtuous and transparent” government should want to identify leakers, with others arguing some secrecy and anonymous sources are essential to accountability.
Environmental and resource costs
- Paper reportedly estimates
5000 GPU-days (120k GPU-hours) and ~20 tons CO₂-eq for all experiments. - Some compare this to emissions from dozens of long flights; others note this is small relative to large LLM training runs but still nontrivial.
- Discussion around limitations of such metrics: variable energy sources, ignored embodied hardware emissions, human activity, and downstream usage.
- Tension between doing impactful research vs. its environmental cost; some favor transparency and awareness over abstention.
Text vs. image watermarking and steganography
- One commenter hoped for text watermarking; others argue text is too easily edited (e.g., postprocessing that strips simple patterns).
- Counterpoint: probabilistic text watermarking schemes can bias token selection so that generated text statistically reveals origin, though they remain vulnerable to heavy editing.
- Clarification that “invisible watermarking,” “steganography,” and “information hiding” are related but distinct: watermarking seeks robust, persistent linkage to content; steganography focuses on undetectability.
- Side debate over whether such techniques amount to “security by obscurity,” with pushback emphasizing the difference between secret algorithms and secret keys.
Ecosystem and deployment considerations
- Some predict large platforms (e.g., major social networks) can unilaterally standardize watermarking, undercutting smaller “authenticity badge” startups.
- Others note parallels to long-standing DRM and fear legal/DMCA-style enforcement around watermarks.
- Minor point that quantizing model weights might break embedded watermark modules, though the intended deployment is mostly at the API/service layer.