Finetune LLMs to teach them ANYTHING with Huggingface and Pytorch | Step-by-step tutorial

แชร์
ฝัง
  • เผยแพร่เมื่อ 1 ธ.ค. 2024

ความคิดเห็น • 29

  • @prashlovessamosa
    @prashlovessamosa หลายเดือนก่อน +6

    This is so top notch education great thanks 🙏👍

    • @avb_fj
      @avb_fj  หลายเดือนก่อน

      Thanks! Glad you enjoyed it!

  • @saimahmed1095
    @saimahmed1095 25 วันที่ผ่านมา

    just by seeing title make me to subscribe. Really needed Thanks bhai.

    • @avb_fj
      @avb_fj  20 วันที่ผ่านมา

      Haha you are welcome bhai! Glad to have you as a sub. :)

  • @JaysThoughts-q5e
    @JaysThoughts-q5e หลายเดือนก่อน +1

    Fun video. You explained it well.

  • @zeelthumar
    @zeelthumar หลายเดือนก่อน +1

    This is crazyyy man keep it up

    • @avb_fj
      @avb_fj  หลายเดือนก่อน

      Thanks!

  • @kenchang3456
    @kenchang3456 หลายเดือนก่อน

    Great consumable explanation, thanks. Nice change of glasses BTW.

    • @avb_fj
      @avb_fj  หลายเดือนก่อน +1

      Haha thanks for noticing the glasses! I totally did not accidentally sit on my old pair.

  • @PedroGabriel-jr3rb
    @PedroGabriel-jr3rb หลายเดือนก่อน

    Invaluable content, thanks!

  • @mbpiku
    @mbpiku หลายเดือนก่อน

    Thanks. Great video.

  • @boogati9221
    @boogati9221 หลายเดือนก่อน +1

    Hey AVB, great video! Do you have plans on making a video for finetuning embeddings? I believe this topic is extremely useful for custom RAG pipelines.

    • @avb_fj
      @avb_fj  หลายเดือนก่อน +2

      Great idea. I didn’t have plans to make this, but sounds like a super important topic. I’ll add it to my TODO list!

    • @boogati9221
      @boogati9221 29 วันที่ผ่านมา

      Legit

  • @sukantdebnath4463
    @sukantdebnath4463 11 วันที่ผ่านมา

    Can you share the notebook as open source pls?

  • @Urvil_Panchal
    @Urvil_Panchal หลายเดือนก่อน +1

    The content is Top Notch ❤‍🔥 , easy to understand. What's your linkedin bro?

  • @ashmin.bhattarai
    @ashmin.bhattarai หลายเดือนก่อน +1

    What is the system spec (RAM / VRAM / GPU) you have used for fine tuning 1B Model?

    • @avb_fj
      @avb_fj  หลายเดือนก่อน +3

      Great question. Something I should have mentioned in the video. This is MacBook Pro M2 16GB ram.
      If you have CUDA gpus, you could leverage better quantization. For my machine, I was able to train with batch size of 8 in float32… the sequence lengths were around 250 on average for this task.
      Honestly, if I were working on a real project I’ll rent GPU on the cloud and train there after prototyping locally. Since this is a YT video and it’s for education, I decided to not go into cloud servers.

    • @ashmin.bhattarai
      @ashmin.bhattarai หลายเดือนก่อน

      @@avb_fj If so, one could easy replicate this exact project on free colab or kaggle..

  • @PaneerPakoda-d9e
    @PaneerPakoda-d9e หลายเดือนก่อน +1

    Hello sir how to actually create custom tokenizers for like let's say Odia language?

    • @avb_fj
      @avb_fj  หลายเดือนก่อน +1

      I would check if any of the existing open source LLMs have Oriya words/alphabets in them. If not, I’d look at things like “Byte pair encoding” and “Subword tokenizers”. Huggingface has a Tokenizers library that provides useful APIs for these.
      If you find a pretrained model where a fair amount of Odia words are in the vocab then that’ll be pretty great. If they are not present, then the job becomes quite difficult, coz you’d either need to add new vocab tokens into existing models, or train your own Odia LM from scratch.

    • @PaneerPakoda-d9e
      @PaneerPakoda-d9e หลายเดือนก่อน

      @@avb_fj thanks for the info🙏🏻

  • @ratand4619
    @ratand4619 หลายเดือนก่อน

    damn i just came across this video and ive subscribed already. I come from arts background but ai has got all my focus lately, and i believe your channel would be the greatest help for me and people like me. Just one view, if you could explain technical points with the simplest explaination taking simplest examples, it'll be of a great help for people like me who don't come from this background. Thousands of people woulf benefit from your amazing content.

    • @avb_fj
      @avb_fj  หลายเดือนก่อน

      Thank you so much for your kind words! Welcome to the channel! Regarding your feedback - I think because this topic was pretty intense, I tried to keep the technical points as concise as possible. I already have some older videos that describe attention, transformers, and LLMs, in more intuitive ways, so I opted to point the viewers towards those videos depending on what they are looking for. That said, I am definitely still working on my skills of distilling down technical points to it's simplest forms, and I hope to keep on improving one video at a time!

  • @mohsenghafari7652
    @mohsenghafari7652 หลายเดือนก่อน

    thanks . please share code.

    • @avb_fj
      @avb_fj  หลายเดือนก่อน +1

      As mentioned in the video, all the code produced in this video and all other videos in the channel are shared on my Patreon.
      www.patreon.com/c/NeuralBreakdownwithAVB

  • @ErkanER1227
    @ErkanER1227 หลายเดือนก่อน

    I wish we didn't have to pay to access the materials 😢

    • @avb_fj
      @avb_fj  หลายเดือนก่อน

      Email me at neural.avb@gmail.com