Transferable Multi-Bit Watermarking Across Frozen Diffusion Models via Latent Consistency Bridges

📅 2026-03-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work proposes DiffMark, a novel watermarking method that overcomes the limitations of existing approaches in frozen diffusion models by simultaneously achieving high multi-bit capacity, cross-model transferability, and efficient detection without relying on time-consuming inversion or model-specific fine-tuning. DiffMark injects learnable perturbations at each denoising step to accumulate watermark signals in the latent space, enabling multi-bit watermark detection with a single forward pass and supporting image-level dynamic keys. By leveraging a Latent Consistency Model (LCM) as a differentiable training proxy, the method reduces backpropagation steps from 50 to just 4, dramatically improving computational efficiency. The approach requires no retraining for new models, achieves detection in 16.4 milliseconds—45× faster than prior methods—and demonstrates strong robustness against distortions, regeneration, and adversarial attacks.

Technology Category

Application Category

📝 Abstract
As diffusion models (DMs) enable photorealistic image generation at unprecedented scale, watermarking techniques have become essential for provenance establishment and accountability. Existing methods face challenges: sampling-based approaches operate on frozen models but require costly $N$-step Denoising Diffusion Implicit Models (DDIM) inversion (typically N=50) for zero-bit-only detection; fine-tuning-based methods achieve fast multi-bit extraction but couple the watermark to a specific model checkpoint, requiring retraining for each architecture. We propose DiffMark, a plug-and-play watermarking method that offers three key advantages over existing approaches: single-pass multi-bit detection, per-image key flexibility, and cross-model transferability. Rather than encoding the watermark into the initial noise vector, DiffMark injects a persistent learned perturbation $δ$ at every denoising step of a completely frozen DM. The watermark signal accumulates in the final denoised latent $z_0$ and is recovered in a single forward pass. The central challenge of backpropagating gradients through a frozen UNet without traversing the full denoising chain is addressed by employing Latent Consistency Models (LCM) as a differentiable training bridge. This reduces the number of gradient steps from 50 DDIM to 4 LCM and enables a single-pass detection at 16.4 ms, a 45x speedup over sampling-based methods. Moreover, by this design, the encoder learns to map any runtime secret to a unique perturbation at inference time, providing genuine per-image key flexibility and transferability to unseen diffusion-based architectures without per-model fine-tuning. Although achieving these advantages, DiffMark also maintains competitive watermark robustness against distortion, regeneration, and adversarial attacks.
Problem

Research questions and friction points this paper is trying to address.

watermarking
diffusion models
multi-bit
transferability
frozen models
Innovation

Methods, ideas, or system contributions that make the work stand out.

multi-bit watermarking
frozen diffusion models
latent consistency models
cross-model transferability
single-pass detection
🔎 Similar Papers
No similar papers found.
H
Hong-Hanh Nguyen-Le
University College Dublin, Ireland
V
Van-Tuan Tran
Trinity College Dublin, Ireland
T
Thuc D. Nguyen
Ho Chi Minh City University of Science
Nhien-An Le-Khac
Nhien-An Le-Khac
Associate Professor of Digital Forensics and Cyber Security, University College Dublin
Digital ForensicsCybersecurityAI SecurityAI ForensicsKnowledge Engineering