Deterministic Image Editing with DDPM Inversion, DDIM Inversion, Null Inversion and Prompt-to-Prompt

แชร์
ฝัง
  • เผยแพร่เมื่อ 10 พ.ย. 2024

ความคิดเห็น • 5

  • @EkShunya
    @EkShunya 3 หลายเดือนก่อน +1

    great one, really liked it
    thanks

  • @ml-ok3xq
    @ml-ok3xq หลายเดือนก่อน

    congrats on writing a paper! i notice that another recent paper from NVIDIA uses a unit vector for attention (nGPT) where the dot product is naturally equal to cosine as the lengths are one. are these two works related to each other in any way?

    • @gabrielmongaras
      @gabrielmongaras  หลายเดือนก่อน +1

      Thanks!! I only read through the nGPT paper briefly, but I think nGPT was trying to make softmax attention/transformers more expressive and efficient by changing a few things. They do normalize before they apply the softmax function, making the logits a cosine similarity between -1 and 1. However they keep the softmax operation which forces the model to stay quadratic in terms of complexity. The paper I worked on removed the softmax function which allowed the attention mechanism to be changed into an RNN which is linear in complexity.

  • @陈兆伟-s5w
    @陈兆伟-s5w 3 หลายเดือนก่อน

    How is the equality in DDPM established in 17:49?

    • @gabrielmongaras
      @gabrielmongaras  3 หลายเดือนก่อน +1

      Looks like I forgot to write out the square root over the first term. As for the inner term that got turned into a fraction, I just multiplied sqrt{1-a_t} by the fraction sqrt{1-a_t}/sqrt{1-a_t}.