Visualizing the Latent Space: This video will change how you imagine neural nets!

แชร์
ฝัง
  • เผยแพร่เมื่อ 16 ม.ค. 2025

ความคิดเห็น • 53

  • @balcofono666
    @balcofono666 ปีที่แล้ว +33

    This is the most concise, simplest and best explanation of latent space that I've found so far!

    • @avb_fj
      @avb_fj  ปีที่แล้ว +1

      Thanks a lot! Glad you enjoyed it! 🙏🏽

  • @ManiSaintVictor
    @ManiSaintVictor 10 หลายเดือนก่อน +4

    Excellent clarification of multiple concepts in one pass. Helped me relate the encoders relative to latent space in a much more accessible metaphor. Thank you.

  • @abdellahfazza4534
    @abdellahfazza4534 11 หลายเดือนก่อน +2

    10 min felt like 30 min, I had so many rewinds during the vid. the video is so full of info, thanks a lot.

  • @ArianaSmith-t2x
    @ArianaSmith-t2x หลายเดือนก่อน

    Thank you for the first video that has allowed me to understand the latent space on some intuitional level

  • @matthewli647
    @matthewli647 10 หลายเดือนก่อน +2

    This video is clear and concise, amazing work!

  • @amdenis
    @amdenis 5 หลายเดือนก่อน

    Great job! I have used your videos to help several recent grads, who had some gaps in understanding. You are a very good teacher.

    • @avb_fj
      @avb_fj  5 หลายเดือนก่อน

      Awesome! Thank you!

  • @habilzare2451
    @habilzare2451 5 หลายเดือนก่อน +1

    Thanks Neural Breakdown with AVB! I always wanted to know how "latent space arithmetic" works.

  • @intellectualobserver
    @intellectualobserver 10 หลายเดือนก่อน +1

    This video is so fascinating. Amazing work.

    • @avb_fj
      @avb_fj  10 หลายเดือนก่อน

      Thanks! Glad you enjoyed it!

  • @mohanpugaz
    @mohanpugaz 10 หลายเดือนก่อน +2

    Amazing Explanation
    Thanks a lot!

  • @qwertyuuytrewq825
    @qwertyuuytrewq825 ปีที่แล้ว +1

    Great explanations! Getting better understanding how some parts of Stable Diffusion work without any efforts )

  • @draftbkshen9088
    @draftbkshen9088 10 หลายเดือนก่อน +1

    This video is super cool. It's good to see those visualized concept

    • @avb_fj
      @avb_fj  10 หลายเดือนก่อน

      Thanks!😊

  • @calum.macleod
    @calum.macleod ปีที่แล้ว +4

    A great description of interpreting deep learing models. Well done!

  • @DaeOh
    @DaeOh ปีที่แล้ว +2

    I hope you can fix the quiet audio!

    • @avb_fj
      @avb_fj  ปีที่แล้ว +1

      Unfortunately YT doesn't allow to increase the volume after posting videos. I could re-upload it, but idk if it'll be really worth it. I'll just take this as lesson for future videos, thanks for the comment!

  • @ParthivShah
    @ParthivShah 10 หลายเดือนก่อน +1

    Truly Amazing

  • @FrejLorenzen
    @FrejLorenzen 8 หลายเดือนก่อน

    Which tools did you use to be able to change each principal component and see its effect on output image?

    • @avb_fj
      @avb_fj  8 หลายเดือนก่อน

      If I remember correctly, I just used ipywidgets inside a jupyter notebook to do the UI and display. I also wrote the logic for the PCA (sklearn), encoding/decoding, and interpolating the latent vectors.

  • @any9912
    @any9912 หลายเดือนก่อน

    Great video!

  • @DanielTheCunningham
    @DanielTheCunningham 11 หลายเดือนก่อน

    Excellent video! Thanks for your work!
    QQ: Is there a repo for the real-time image manipulation software you used as your demo?

  • @Shaunmcdonogh-shaunsurfing
    @Shaunmcdonogh-shaunsurfing 9 หลายเดือนก่อน +1

    Excellent video

  • @ricdodo
    @ricdodo หลายเดือนก่อน

    How does the latent space work exactly?? Does it learn just a global representation (like the slime is typically in the center) or also a local representation? I mean, if we perform the same operation with images rotated randomly, do the latent space arithmetics stop working? Do you know of any papers that explore this in depth. thx in advance

  • @JeremyMamisao
    @JeremyMamisao ปีที่แล้ว +1

    Great video, thanks

  • @coder-c6961
    @coder-c6961 ปีที่แล้ว +1

    This was a insanely good video and explanation ty

    • @avb_fj
      @avb_fj  ปีที่แล้ว

      Thanks! Awesome to hear that! 😊

    • @coder-c6961
      @coder-c6961 ปีที่แล้ว +1

      @avb_fj Thank you! I'd like to see one where you code the encoder and decoder I'm coding a autoencoder and it's a little tough trying to find a good balance of reduction but also keeling the important details

  • @qwioqwe
    @qwioqwe ปีที่แล้ว

    Thanks for great video! Very well explained!

  • @karmeshmeartia3784
    @karmeshmeartia3784 10 หลายเดือนก่อน

    can we do same with the pixels to enhance the image?

    • @avb_fj
      @avb_fj  10 หลายเดือนก่อน

      Can you clarify what you meant by “doing the same with pixels”?

  • @shahriarshaon9418
    @shahriarshaon9418 ปีที่แล้ว +2

    I really appreciate your lucid explanation. Superb. I wanna request you if you could enhance the sound quality a bit. Good wishes and thanks for such videos

    • @avb_fj
      @avb_fj  ปีที่แล้ว

      Thanks! I’ll keep that in mind going forward…

  • @rishidixit7939
    @rishidixit7939 2 หลายเดือนก่อน

    I think of it like CNNs were good at Pattern Recognition and Identification so they were well suited for Classification type tasks but in Generation we need more nuanced understanding (more data is needed which cannot be picked as a pattern or is more abstract) and so the concept of Latent Space was introduced.
    Is my thinking in right direction or is it wrong ? Please do tell

    • @avb_fj
      @avb_fj  2 หลายเดือนก่อน +1

      Thanks for the comment!
      It’s a bit more nuanced though coz CNNs have been successfully used for both discriminative tasks (like classification) as well as generative tasks (GANs, Diffusion). At its heart, convolution is a general purpose operation to process spatial information within images. The rest of it's power really comes from special architectural choices, quality and diversity of training data, loss functions, etc. The concept of "Latent Space" has pretty deep roots in Deep Learning - it all probably came from this term called "Representation Learning" or "Manifold Learning", which don't get said often these days. As discussed in the video, the basic idea is that neural networks form their own numerical representations within their weights when we train them on any kind of data, and these representations are often meaningful and interpretable in the right experimental settings.
      I'll suggest to check these two videos out where I tried to explain Convolution and CNNs from the ground up.
      th-cam.com/video/kebSR2Ph7zg/w-d-xo.html
      th-cam.com/video/N_PocrMHWbw/w-d-xo.html

    • @rishidixit7939
      @rishidixit7939 2 หลายเดือนก่อน

      ​@@avb_fj Thanks for this.
      I do think CNNs are good in that but in the case of GANs the Generator Discriminator pair was there to kind of find those hidden or more nuanced features which are more than patterns.
      Seems like both of them (GANs and VAE) were 2 different methods to solve a similar problem of capturing more nuanced features.
      I am just a beginner in this field so I can be completely wrong. But what I have observed is that major part of Deep Learning theory revolves around finding more nuanced features and how they are inter related.

  • @khanhbq
    @khanhbq 9 หลายเดือนก่อน +1

    Great content!

  • @neithanm
    @neithanm ปีที่แล้ว +2

    Oh man. I was a bit lost when you where saying encoder this decoder that but the smile example at 6:50 hit right on the nail. It's indeed mindblowing. I'd love to know more about AI for outsiders, subscribed.
    PS: A concept I picked up from Ezra is that AI turns semantics into geometry. So you can do king - man + woman and get queen! (paraphrasing). If you could expand on this and give more examples in different modalities... that'd be awesome.

    • @avb_fj
      @avb_fj  ปีที่แล้ว +1

      Nice…glad you enjoyed it and stuck around for the whole thing. The semantic example is pretty awesome yeah… I’ve brought it up in the channel in my History of NLP video, but more examples on different modalities seems like a nice idea for a video!

  • @exoticcoder5365
    @exoticcoder5365 ปีที่แล้ว +1

    Thank you sir ! You clear the concept of latent space for me ! And I can’t wait to click on your multi modal video in this channel

  • @w000w00t
    @w000w00t ปีที่แล้ว +1

    Just subscribed after your NeRF video, and this one is awesome too! You, Yannic, and Two Minute Papers are great at making AI content relatable and interesting and freaking cool :) What a time to be alive! lol

    • @avb_fj
      @avb_fj  ปีที่แล้ว +1

      Wow that’s high praise! Those two are definitely an inspiration, so I’m kinda feeling surreal reading this! Thank you so much!! 🙌🏼🙌🏼

  • @TP-ct7qm
    @TP-ct7qm ปีที่แล้ว

    Awesome content!

    • @avb_fj
      @avb_fj  ปีที่แล้ว

      🙏🏽🙌🏼

  • @geoffgodwin
    @geoffgodwin ปีที่แล้ว +4

    This is a fantastic break down. Great pacing, wonderful examples with easy to follow metaphors. Fix your audio and keep em coming!

  • @Korea-Lens
    @Korea-Lens 4 หลายเดือนก่อน +1

    really great overview of topics. subd n liked each im watching!

  • @Atheist-Libertarian
    @Atheist-Libertarian 4 หลายเดือนก่อน

    wow

  • @pkjindal953
    @pkjindal953 4 วันที่ผ่านมา

    Great information. Considerably low voice. Please take care next time.

  • @fokushodl52103
    @fokushodl52103 ปีที่แล้ว +2

    Great content!

  • @teleprint-me
    @teleprint-me ปีที่แล้ว

    Great content! Just as a FYI, might want to turn up the Mic volume. It's easier to lower the volume than to turn it up from the consumers POV.

    • @avb_fj
      @avb_fj  ปีที่แล้ว +1

      Thanks for the feedback! Will keep it in mind for the next one…

  • @Essentia-Channel
    @Essentia-Channel 11 หลายเดือนก่อน

    Great video but the audio level is way too low. Also the video and audio is not in sync.