LoRA - Low-rank Adaption of AI Large Language Models: LoRA and QLoRA Explained Simply

แชร์
ฝัง
  • เผยแพร่เมื่อ 28 ม.ค. 2025

ความคิดเห็น • 42

  • @IntuitiveAndExhaustive
    @IntuitiveAndExhaustive ปีที่แล้ว +45

    This isn't really accurate. LoRA doesn't really make your model smaller, it makes the fine tuning process require fewer resources. As far as I understand the final model size is about the same, if not slightly bigger.

    • @amortalbeing
      @amortalbeing ปีที่แล้ว +2

      he meant during training I guess, but you are right, Lora is for fine-tuning

    • @viixby9481
      @viixby9481 5 หลายเดือนก่อน +2

      A better analogy I assume with the lego thing, would be that, it gives you a little handbook on how to build that specific thing with the legos. Not handpicking certain legos.

    • @sharathkumar8422
      @sharathkumar8422 3 หลายเดือนก่อน

      Final model is slightly bigger as it has the LoRA layers added on top of the original layers unless you removed a few of the original before adding the LoRA ones. The training process however requires far less computational resources cause you're only training the newly added LoRA layers and freezing all the other model layers.

  • @seeess925
    @seeess925 27 วันที่ผ่านมา +2

    This makes no sense. Keeps saying models over and over with no info on what it h means. I get the overall point of being limited. But no info on what it actually is it what it does.. Essentially pretty much know nothing about what lora means

  • @lighthousesaunders7242
    @lighthousesaunders7242 ปีที่แล้ว +4

    Fine tunning?
    Stacking and transfering?
    Thanks for a great video.

  • @MasterBrain182
    @MasterBrain182 ปีที่แล้ว +1

    Astonishing content Man 🚀

  • @DarkDiripti
    @DarkDiripti 10 หลายเดือนก่อน +5

    Only watched the first lego example, and that is just plain wrong. LoRA does not make the model smaller, that analogy does not hold at all. Don't want to know what follows from such a bad analgoy.

  • @El_MA11
    @El_MA11 4 หลายเดือนก่อน

    That was excellent. Nice teaching skills, Wes.

  • @m.e.3614
    @m.e.3614 17 วันที่ผ่านมา

    Thank you for this clear explanation!

  • @Trinad356
    @Trinad356 ปีที่แล้ว +8

    You explanation is really amazing one can easily understand without having any background knowledge, thank you very much.

  • @sheevys
    @sheevys ปีที่แล้ว +2

    Is there a speed improvement during training only or also at inference?

    • @Lampshadx
      @Lampshadx ปีที่แล้ว

      Training only - you may encounter latency issues during inference due to separately loading the base model and the LoRA model, but most libraries will allow you to merge them, so it would end up being equivalent to the original model.

    • @Lampshadx
      @Lampshadx ปีที่แล้ว

      To actually run inference and generate predictions after fine-tuning Lora, you would need to combine the original large base model weight matrices with the small updated factor matrices that Lora learns. So at inference time, you still have essentially the same enormous number of parameters as the original foundation model.
      The key efficiency gains are seen during the adaptation/fine-tuning process. By only updating a tiny fraction of parameters, Lora allows much quicker and cheaper adaptation compared to full fine-tuning.
      But once the adapted model is ready for deployment and inference, multiplying/merging those factor matrices back into the original base weights results in effectively the full set of parameters at inference time.

  • @TheCradmin
    @TheCradmin ปีที่แล้ว +4

    This video needs more love. Thank you man, so well communicated.

  • @happyday.mjohnson
    @happyday.mjohnson ปีที่แล้ว

    i subscribed to your channel after watching your explanation. Thank you for your clarity.

  • @Daligliding
    @Daligliding 3 หลายเดือนก่อน

    What kind of tool is used making the animation?

  • @scottstout
    @scottstout ปีที่แล้ว +1

    Is it possible/reasonable to use LoRA w/ GPT-4?

    • @fusseldieb
      @fusseldieb ปีที่แล้ว +1

      Afaik GPT-4 is closed-source

  • @ArunkumarMTamil
    @ArunkumarMTamil 9 หลายเดือนก่อน

    how is Lora fine-tuning track changes from creating two decomposition matrix? How the ΔW is determined?

  • @AurobindoTripathy
    @AurobindoTripathy ปีที่แล้ว

    all that images-filling by scribbling inside the lines? How does that support your content (which in fine)?

  • @moeinhasani8718
    @moeinhasani8718 ปีที่แล้ว

    really good video for a high-level understanding of the concept. I wish there was a little bit of math included as well, just very high level mentioning what mathematical steps are taken.

  • @ntesla66
    @ntesla66 ปีที่แล้ว +2

    1. a large beer cask. 2. a measure of capacity, usually equal to 252 wine gallons. Otherwise well done!

  • @sergeibogdanov572
    @sergeibogdanov572 ปีที่แล้ว

    Hello, what is the name of the software you use to draw?

  • @chyldstudios
    @chyldstudios ปีที่แล้ว +1

    nice explainer

  • @pongtrometer
    @pongtrometer ปีที่แล้ว

    I’m trying to follow your recommendation of learning python, I’m not a programmer whatsoever, so just from watching this overview, which is great by the way. Is there LoRA-esque way of learning python , so that I can be creative ,as I learn python. Just like using LoRA’s in SDXL, in combinations to create new image recipes. I hope to learn python so that I can get involved with making LoRA’s for sound design, not grounding breaking, but definitely enabling sound designers to create with new sonic colours. Any advice would be much appreciated. Thanks in advance Wes / comments community.

  • @KentPage-j8u
    @KentPage-j8u 4 หลายเดือนก่อน

    518 Skiles Shores

  • @FrankWilliams-r4b
    @FrankWilliams-r4b 3 หลายเดือนก่อน

    Satterfield Hill

  • @ceciliaedwards552
    @ceciliaedwards552 3 หลายเดือนก่อน

    204 Oberbrunner Valleys

  • @ChristyLynn-j1t
    @ChristyLynn-j1t 4 หลายเดือนก่อน

    458 McGlynn Circle

  • @HansenHugo
    @HansenHugo 3 หลายเดือนก่อน

    50849 Klocko Plaza

  • @AnneMorales-y2r
    @AnneMorales-y2r 4 หลายเดือนก่อน

    Hal Crescent

  • @SherillBriner
    @SherillBriner 4 หลายเดือนก่อน

    290 Block Place

  • @iainmackenzie1995
    @iainmackenzie1995 3 หลายเดือนก่อน

    Pieces of Lego are not called Legos

  • @GeorgeHarris-f8r
    @GeorgeHarris-f8r 4 หลายเดือนก่อน

    Elisabeth Field

  • @robertputneydrake
    @robertputneydrake ปีที่แล้ว +1

    *powerful not powerfull :)

    • @Marcus_Berger1701
      @Marcus_Berger1701 10 หลายเดือนก่อน

      Unliimmmmmited powaaaaaafulllllll 😁

  • @CynthiaEvangelista-b8h
    @CynthiaEvangelista-b8h 4 หลายเดือนก่อน

    Jerome Groves

  • @cgqqqq
    @cgqqqq 4 หลายเดือนก่อน

    useless video, you can make like ~10 videos by just replacing LORA with anyother AI jargons

  • @halilxxx
    @halilxxx 2 หลายเดือนก่อน

    its not true

  • @aaronsayeb6566
    @aaronsayeb6566 5 หลายเดือนก่อน

    just wasted 4 minutes on this