kexun
    • Create new note
    • Create a note from template
      • Sharing URL Link copied
      • /edit
      • View mode
        • Edit mode
        • View mode
        • Book mode
        • Slide mode
        Edit mode View mode Book mode Slide mode
      • Customize slides
      • Note Permission
      • Read
        • Only me
        • Signed-in users
        • Everyone
        Only me Signed-in users Everyone
      • Write
        • Only me
        • Signed-in users
        • Everyone
        Only me Signed-in users Everyone
      • Engagement control Commenting, Suggest edit, Emoji Reply
    • Invite by email
      Invitee

      This note has no invitees

    • Publish Note

      Share your work with the world Congratulations! 🎉 Your note is out in the world Publish Note

      Your note will be visible on your profile and discoverable by anyone.
      Your note is now live.
      This note is visible on your profile and discoverable online.
      Everyone on the web can find and read all notes of this public team.
      See published notes
      Unpublish note
      Please check the box to agree to the Community Guidelines.
      View profile
    • Commenting
      Permission
      Disabled Forbidden Owners Signed-in users Everyone
    • Enable
    • Permission
      • Forbidden
      • Owners
      • Signed-in users
      • Everyone
    • Suggest edit
      Permission
      Disabled Forbidden Owners Signed-in users Everyone
    • Enable
    • Permission
      • Forbidden
      • Owners
      • Signed-in users
    • Emoji Reply
    • Enable
    • Versions and GitHub Sync
    • Note settings
    • Note Insights New
    • Engagement control
    • Transfer ownership
    • Delete this note
    • Save as template
    • Insert from template
    • Import from
      • Dropbox
      • Google Drive
      • Gist
      • Clipboard
    • Export to
      • Dropbox
      • Google Drive
      • Gist
    • Download
      • Markdown
      • HTML
      • Raw HTML
Menu Note settings Note Insights Versions and GitHub Sync Sharing URL Create Help
Create Create new note Create a note from template
Menu
Options
Engagement control Transfer ownership Delete this note
Import from
Dropbox Google Drive Gist Clipboard
Export to
Dropbox Google Drive Gist
Download
Markdown HTML Raw HTML
Back
Sharing URL Link copied
/edit
View mode
  • Edit mode
  • View mode
  • Book mode
  • Slide mode
Edit mode View mode Book mode Slide mode
Customize slides
Note Permission
Read
Only me
  • Only me
  • Signed-in users
  • Everyone
Only me Signed-in users Everyone
Write
Only me
  • Only me
  • Signed-in users
  • Everyone
Only me Signed-in users Everyone
Engagement control Commenting, Suggest edit, Emoji Reply
  • Invite by email
    Invitee

    This note has no invitees

  • Publish Note

    Share your work with the world Congratulations! 🎉 Your note is out in the world Publish Note

    Your note will be visible on your profile and discoverable by anyone.
    Your note is now live.
    This note is visible on your profile and discoverable online.
    Everyone on the web can find and read all notes of this public team.
    See published notes
    Unpublish note
    Please check the box to agree to the Community Guidelines.
    View profile
    Engagement control
    Commenting
    Permission
    Disabled Forbidden Owners Signed-in users Everyone
    Enable
    Permission
    • Forbidden
    • Owners
    • Signed-in users
    • Everyone
    Suggest edit
    Permission
    Disabled Forbidden Owners Signed-in users Everyone
    Enable
    Permission
    • Forbidden
    • Owners
    • Signed-in users
    Emoji Reply
    Enable
    Import from Dropbox Google Drive Gist Clipboard
       Owned this note    Owned this note      
    Published Linked with GitHub
    • Any changes
      Be notified of any changes
    • Mention me
      Be notified of mention me
    • Unsubscribe
    ## General Response Regarding the Robustness of StegaStamp Our experiments involving StegaStamp were conducted under a constrained noise scale, as elaborated in the "Proposed Attacks" paragraph in the Evaluation section. Specifically, the noise scale was set to 60 steps across all evaluated watermarking techniques, including StegaStamp. This was to maintain a consistent utility loss across different watermarks. StegaStamp's relatively poor visual quality and high perturbation, as documented in Table 2 (especially in terms of PSNR and FID) and Figure 3 of our manuscript, render it more resilient to watermark removal when constrained by a fixed, moderate noise scale. However, our new experiments suggest that increasing the noise scale in diffusion models significantly compromises StegaStamp's robustness. Below are the results illustrating this effect: | [Steps] | [Noise level] | [Avg bit acc] | [Avg detect acc] | | ----- | ----------- | ----------- | -------------- | | 60 | 0.251 | 0.861 | 0.991 | | 150 | 0.457 | 0.709 | 0.861 | | 200 | 0.571 | 0.658 | 0.677 | | 250 | 0.696 | 0.614 | 0.405 | | 300 | 0.832 | 0.585 | 0.229 | | 350 | 0.988 | 0.558 | 0.087 | | 400 | 1.164 | 0.546 | 0.062 | ## Reviewer A **R1. On the Robustness of StegaStamp** We kindly refer you to our general response. **R2. Recent Work in Watermarking** We thank the reviewer for bringing to our attention this concurrent work WOUAF [1]. Upon your suggestion, we have attempted to evaluate its resilience to our attack; however, its code is not publicly available as of now. We have initiated correspondence with the authors of WOUAF. We will incorporate its evaluation in the revision. ## Reviewer B **R1. Trade-off Curve for StegaStamp** Thank you for your suggestion! In our paper, we show this trade-off curve against the DwtDctSvd watermark (Figure 6). Upon your request, we have generated a table for StegaStamp that exhibits the trade-off under different noise scales. We kindly refer you to the table in our general response. **R2. Quantifying Utility Loss for Tree-Ring** Indeed, the preservation of semantic meaning is not easily measured with conventional metrics like L2. In the original Tree-Ring paper, the authors have employed Frechet Inception Distance (FID) [2] to assess generation quality, while also leveraging the CLIP score [3] to ascertain semantic consistency using OpenCLIP-ViT/G. Alternative metrics such as the BLIP score [4] and ImageReward [5] are also available for assessing vision-language alignment. In light of your suggestion, we will discuss and evaluate more on semantic utility loss quantification in the revision. ## Reviewer C **R1. > "Guarantee similar to differential privacy"** The reviewer is right that we used ideas and techniques from differential privacy. Our technical contribution is a novel application of modern techniques (e.g., f-DP and GDP) from the DP literature to a new problem. Let us summarize the main differences: - f-CWF is different from DP (and f-DP) because f can depend on individual instance (image, watermark) pair. The distinction is important for us to quantify the effect of the embedding phi via the "Local Lipschitz" property. - The utility bound -- Theorem 4.8 is new and somewhat clean. - The use of modern technique (which is the reason why the presentation seems "contrived") is needed for us to get the tight characterization of the impossibility region as in Figure 4. If we use the classical (eps, delta)-DP then the certified region would be much smaller (and less valuable in practice). - Unlike in most DP mechanisms, we do not need to artificially inject new noise. The noise is inherent to the diffusion model. **R2. > "Adding Gaussian noise was proposed before"** The reviewer is right. Different from existing works, we proposed to add Gaussian noise in the "embedding space" rather than the raw pixel space. Existing work also does not have formal guarantees for watermark removal as we do. **R3. Presentation Issues** We will strive to address all highlighted issues in the next revision. ## Reviewer D **R1. > "Table 5 does not report L2" distance** The PSNR that we reported is calculated using the L2 distance. PSNR = 10* log_10(Num of pixels / L2-dist^2). **R2. > "utility of the theory"** The advantage of the theory is that it is **future-proof**. Our results cover all future watermarks and detection methods, not just those that we empirically evaluated. Figure 4 clearly demonstrates the utility of our theory and its relevance to practice. **R3. StegaStamp and StableSignature** We think we understand the question you raised about interpreting our empirical results regarding “StegaStamp” and “StableSignature” and the extent to which they justify our claim "all invisible watermarks are removable" and why it appears to be "contradictory". Let us address this concern by making the following arguments. First, our results describe a fundamental tradeoff between the “distortion” introduced by the watermark (measured in L2-distance) and the “security” (against removal attacks). It is not a binary predicate, but a continuum of pareto-front. Specifically, if the watermark perturbs the image with a higher L2 distance (as StegaStamp and Stable Signature did), our attack will require adding larger noise to achieve the same level of “watermark-free”ness — and certainly, the reconstructed image will be less similar to the original. Figure 8 in our paper clearly demonstrates that StegaStamp and StableSignature add substantially more distortions in L2 distance than others. The amount of noise we add in the attack, however, is the same across all methods. For that reason, it is not surprising that StegaStamp detector is more effective than others (as in Table 1). To convince you that this is the case, we added experiments with a variety of different noise levels (pasted in our general response). Notably, when the number of steps of Diffusion is set to 400, the detection accuracy is merely 0.062. **R4. > "StegaStamp give SSIM of 0.91 while "our reconstructed image has SSIM of 0.7". "How can you claim that StegaStamp suffers more visual artifacts, but the reconstructed images are good?"** Note that we did not claim that the reconstructed image will be closer than the watermarked image to the original image. In fact, it is expected that the reconstructed image will be more different than the watermarked image when compared to the original — that is the price to pay for removing watermarks. Exceptions are when certain watermarks are “noise-like” (e.g., high-frequency noise), only then we can hope that the attacked images are closer to the original. This may appear to be at odds with our Theorem 4.8. But notice that Theorem 4.8 is a relevant guarantee. It proved that if the original image + noise can be effectively denoised, then the watermarked image + noise can be denoised *almost just as effectively*. Our hunch is that to remove the slightly larger watermark like StegaStamp, we need to introduce larger noise, therefore the baseline — original image + noise — after the denoising / reconstruction by diffusion, will be different from the original image (e.g., in the range of SSIM = 0.7). Lastly, we want to emphasize that being different from the original image does not necessarily mean lower-visual quality (thanks to the ability of Stable-Diffusion to “hallucinate” details). Even if SSIM = 0.7, the reconstructed image can still be quite visually appealing (and semantically similar to the original image). We demonstrate this with more examples in the attachment. **R5. > why "attack should increase both FPs and FNs." "only FNs matter."** Our result covers all detectors, not just those designed for the watermark. That includes the trivial detector that always outputs “Yes, watermarked!”, hence FN = 0 but not useful. ### Reference [1] Kim, Changhoon, et al. "WOUAF: Weight Modulation for User Attribution and Fingerprinting in Text-to-Image Diffusion Models." arXiv preprint arXiv:2306.04744 (2023). [2] Heusel, Martin, et al. "Gans trained by a two time-scale update rule converge to a local nash equilibrium." Advances in neural information processing systems 30 (2017). [3] Radford, Alec, et al. "Learning transferable visual models from natural language supervision." International conference on machine learning. PMLR, 2021. [4] Li, Junnan, et al. "Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation." International Conference on Machine Learning. PMLR, 2022. [5] Xu, Jiazheng, et al. "Imagereward: Learning and evaluating human preferences for text-to-image generation." arXiv preprint arXiv:2304.05977 (2023).

    Import from clipboard

    Paste your markdown or webpage here...

    Advanced permission required

    Your current role can only read. Ask the system administrator to acquire write and comment permission.

    This team is disabled

    Sorry, this team is disabled. You can't edit this note.

    This note is locked

    Sorry, only owner can edit this note.

    Reach the limit

    Sorry, you've reached the max length this note can be.
    Please reduce the content or divide it to more notes, thank you!

    Import from Gist

    Import from Snippet

    or

    Export to Snippet

    Are you sure?

    Do you really want to delete this note?
    All users will lose their connection.

    Create a note from template

    Create a note from template

    Oops...
    This template has been removed or transferred.
    Upgrade
    All
    • All
    • Team
    No template.

    Create a template

    Upgrade

    Delete template

    Do you really want to delete this template?
    Turn this template into a regular note and keep its content, versions, and comments.

    This page need refresh

    You have an incompatible client version.
    Refresh to update.
    New version available!
    See releases notes here
    Refresh to enjoy new features.
    Your user state has changed.
    Refresh to load new user state.

    Sign in

    Forgot password

    or

    By clicking below, you agree to our terms of service.

    Sign in via Facebook Sign in via Twitter Sign in via GitHub Sign in via Dropbox Sign in with Wallet
    Wallet ( )
    Connect another wallet

    New to HackMD? Sign up

    Help

    • English
    • 中文
    • Français
    • Deutsch
    • 日本語
    • Español
    • Català
    • Ελληνικά
    • Português
    • italiano
    • Türkçe
    • Русский
    • Nederlands
    • hrvatski jezik
    • język polski
    • Українська
    • हिन्दी
    • svenska
    • Esperanto
    • dansk

    Documents

    Help & Tutorial

    How to use Book mode

    Slide Example

    API Docs

    Edit in VSCode

    Install browser extension

    Contacts

    Feedback

    Discord

    Send us email

    Resources

    Releases

    Pricing

    Blog

    Policy

    Terms

    Privacy

    Cheatsheet

    Syntax Example Reference
    # Header Header 基本排版
    - Unordered List
    • Unordered List
    1. Ordered List
    1. Ordered List
    - [ ] Todo List
    • Todo List
    > Blockquote
    Blockquote
    **Bold font** Bold font
    *Italics font* Italics font
    ~~Strikethrough~~ Strikethrough
    19^th^ 19th
    H~2~O H2O
    ++Inserted text++ Inserted text
    ==Marked text== Marked text
    [link text](https:// "title") Link
    ![image alt](https:// "title") Image
    `Code` Code 在筆記中貼入程式碼
    ```javascript
    var i = 0;
    ```
    var i = 0;
    :smile: :smile: Emoji list
    {%youtube youtube_id %} Externals
    $L^aT_eX$ LaTeX
    :::info
    This is a alert area.
    :::

    This is a alert area.

    Versions and GitHub Sync
    Get Full History Access

    • Edit version name
    • Delete

    revision author avatar     named on  

    More Less

    Note content is identical to the latest version.
    Compare
      Choose a version
      No search result
      Version not found
    Sign in to link this note to GitHub
    Learn more
    This note is not linked with GitHub
     

    Feedback

    Submission failed, please try again

    Thanks for your support.

    On a scale of 0-10, how likely is it that you would recommend HackMD to your friends, family or business associates?

    Please give us some advice and help us improve HackMD.

     

    Thanks for your feedback

    Remove version name

    Do you want to remove this version name and description?

    Transfer ownership

    Transfer to
      Warning: is a public team. If you transfer note to this team, everyone on the web can find and read this note.

        Link with GitHub

        Please authorize HackMD on GitHub
        • Please sign in to GitHub and install the HackMD app on your GitHub repo.
        • HackMD links with GitHub through a GitHub App. You can choose which repo to install our App.
        Learn more  Sign in to GitHub

        Push the note to GitHub Push to GitHub Pull a file from GitHub

          Authorize again
         

        Choose which file to push to

        Select repo
        Refresh Authorize more repos
        Select branch
        Select file
        Select branch
        Choose version(s) to push
        • Save a new version and push
        • Choose from existing versions
        Include title and tags
        Available push count

        Pull from GitHub

         
        File from GitHub
        File from HackMD

        GitHub Link Settings

        File linked

        Linked by
        File path
        Last synced branch
        Available push count

        Danger Zone

        Unlink
        You will no longer receive notification when GitHub file changes after unlink.

        Syncing

        Push failed

        Push successfully