literally i've searched so many articles blogs and youtube videos to understand transformer from the core basic. And this video come up with all the stuffs , and i must say best video everr!
Best video to learn about transformers. Since you are going through same theories again and again more I heard them more I understood them. Thank you very much.
Hey Ajay, I've watched numerous videos on transformers, but yours stands out as the best-super clear and easy to grasp! Your explanations are fantastic. Could you consider covering transformers for time series in your future videos? That'd be incredibly helpful! Thanks for the great content!
The Best explanation on transfer with code, even tough I am a AI developer and know these things pretty well I watched the full video Love from Banglore
27:40 isn't the key, query and value vectors computed using weight matrices for each of them which have learnable parameters. The input vector is multiplied by weight matrix corresponding to each Q, V and K.
Congrats for 100k.I have been following your channel for about 1 and half years now.Keep up the Good work.I was also asking for some advice.I did andrew ng ML,DL course,GANS and NLP course all from coursera.I implemented about 30 papers on convnets.about 10 on nlp and 15 in GANS .I realized there is a gap in knowledge as the papers can get into proofs which are math or stats heavy which can become hard to understand.i also followed the Karpathy nlp playlist which was great but now i feel like i hit a ceiling.I want to eventually become a MLE (currently a civil eng student).I have gone back to studying DSA because i dont know a recommended way of studying ML going forward.There are topics like optimization theory,game theory,info theory which have been recommended for study but i havent started because am in the middle of my semester.What do you recommend i do or what resources do you recommend i use.How I can reach out privately to talk more on this.
you are currently deep down in the tutorial hell. if you go on you'll never get out as theres always something you dont know. you are at a very good place right now, just start applying for a role you are intrested. you do not need to learn everything, you do not need to know everything. its pointless. noone does that. just focus on one subject/role and try to organize your time and resources to be the best for that role. game theory, information theory, etc are not really needed unless you intend on doing research on the domains involved. if thats the case, your best bet is to enroll in a Phd position and that should guide you properly. if not dont waste your time anymore. apply for a job and learn as you go on. Best of Luck
Appreciate the time you put into this! Refined my understanding of transformer models. I like this graphic you have at 1:19:00 do you have a file of that available somewhere?
When creating a copy for the residual, wouldn't it be necessary to have residual_x = x.clone()? Otherwise wouldn't it just be a reference copy? Also there is a permute statement used to swap the heads and sentences, but I didn't see a second permute statement to swap them back in the returned values tensor before reshaping.
It’s a bit funny that you speak perfect English with absolutely no accent, but you pronounce “matrix” wrong. It should be pronounced “may tricks”, not “mah tricks”.
at 37:42, shouldn't we permute the values to (1, 4, 8, 64) first and then reshape to (1,4,8*64) to accurately concatenate multiple heads for corresponding sentences?
This was the best playlist on TH-cam on Transformers
These complements I do not deserve. Thanks a ton for it :)
Truee
literally i've searched so many articles blogs and youtube videos to understand transformer from the core basic. And this video come up with all the stuffs , and i must say best video everr!
Congrats Ajay! Very well deserved 🎉
Thanks a ton Ritvik! It’s been a long road, but here we are! :)
Best video to learn about transformers. Since you are going through same theories again and again more I heard them more I understood them. Thank you very much.
Hey Ajay, I've watched numerous videos on transformers, but yours stands out as the best-super clear and easy to grasp! Your explanations are fantastic. Could you consider covering transformers for time series in your future videos? That'd be incredibly helpful! Thanks for the great content!
thank you so much ajay, your channel is one of the best in youtube for learning ml
Onwards and upwards to you too Ajay, great vid mate, I actually watched the whole thing
Thank you so much ! I really appreciate your support on this :)
I would have never understood that topic without your video. Thank you.
The Best explanation on transfer with code, even tough I am a AI developer and know these things pretty well I watched the full video
Love from Banglore
Thanks so much for the kind words
This is the best playlist ever. Thank you.
Been following you for a while. Congrats! You definitely made my NLP journey more enjoyable with your informative and entertaining videos.
Thanks so much for the support! And for your kind words! There is definitely more to come :)
Amazing content, you made NLP naturally processable for me.
Thank you for sharing your knowledge with us. This is pure gold !
the best transformer video i have watched
Excellent Video Sir. Respect.
I think at this point I represent a heavy percentage of your views. Great work bro.
best video on transformers ever. even better than Andrej Karpathy's video. you are awesome bro. keep up the great work.
congrats for 100k
Thank you so much for the support :)
27:40 isn't the key, query and value vectors computed using weight matrices for each of them which have learnable parameters. The input vector is multiplied by weight matrix corresponding to each Q, V and K.
I have seen so many transformers videos but this one is outstanding, I also want to request you to make a video on vision transformers too❤
This is incredible and well illustrated. Thanks for putting it together.
Thanks so much for the compliments!
@@CodeEmporium certainly!
omg, you're deserve more and more subscribers
woooaw this is a gift! Thank you!!
Very powerful and great Content
Congrats CE , keep going
Roger! I shall keep making content. Thank you for your support!!
Congrats Ajay bahiya.
Thanks a ton for the support! :D
Thanks for this great video!
you are the best on the youtube for the transformer all the best
Congrats bro
Thank you so much for the support:)
thank you so much
You are very welcome! Thank you for the support:)
Thanks a lot !
Very clear, much appreciated
Super thank you to you. Thanks for commenting too!
That's great man..i am currently studying this...❤
Congratulations!
Thanks a ton for the support
thank you, great job
Congrats for 100k.I have been following your channel for about 1 and half years now.Keep up the Good work.I was also asking for some advice.I did andrew ng ML,DL course,GANS and NLP course all from coursera.I implemented about 30 papers on convnets.about 10 on nlp and 15 in GANS .I realized there is a gap in knowledge as the papers can get into proofs which are math or stats heavy which can become hard to understand.i also followed the Karpathy nlp playlist which was great but now i feel like i hit a ceiling.I want to eventually become a MLE (currently a civil eng student).I have gone back to studying DSA because i dont know a recommended way of studying ML going forward.There are topics like optimization theory,game theory,info theory which have been recommended for study but i havent started because am in the middle of my semester.What do you recommend i do or what resources do you recommend i use.How I can reach out privately to talk more on this.
you are currently deep down in the tutorial hell. if you go on you'll never get out as theres always something you dont know.
you are at a very good place right now, just start applying for a role you are intrested. you do not need to learn everything,
you do not need to know everything. its pointless. noone does that. just focus on one subject/role and try to organize your time and resources
to be the best for that role.
game theory, information theory, etc are not really needed unless you intend on doing research on the domains involved.
if thats the case, your best bet is to enroll in a Phd position and that should guide you properly. if not dont waste your time anymore.
apply for a job and learn as you go on.
Best of Luck
@@amortalbeing Thanks for the advice ,I really appreciate it.
good video, so simiplified
ತುಂಬಾ ಚೆನ್ನಾಗಿದೆ ಹೇಳಿಕೊಡುತ್ತೀರ.
Thanks so much!!! :D
Congrats
Thank you so much!
I'm starting the project to detect stock manipulation using this Transformer Model Your channel is super great! @@CodeEmporium
fuck this was complex but fun
great content AND handsome presenter
Good job man! Thanks a lot. Why do not you make some videos about transformer in time series to predict, classify, ... base on video?
Wow, good job!
Very useful video
Danke!
Need a video on fine tuning the BERT model..
Happy to see you read kannada
Thank you so much, your explanation was amazing. can you please explain how we can use the transformer(encoder-decoder) in time series?
Ooo a video idea. I shall look into this. Thank you for the support regardless :)
Awesome ❤
Best explanation anna love from Andhra Pradesh 😍 😍
Thanks a ton for the kind words :)
@@CodeEmporium 🙂
Nice. Can you bring us an intro to Retentive Networks next?
Appreciate the time you put into this! Refined my understanding of transformer models. I like this graphic you have at 1:19:00 do you have a file of that available somewhere?
nice video
nice work, explanation. appreciate your efforts
would you mind sharing the writing device you are using.
Hey Code Emporium, could you please do a video of how you read & implement ML papers.
When creating a copy for the residual, wouldn't it be necessary to have residual_x = x.clone()? Otherwise wouldn't it just be a reference copy? Also there is a permute statement used to swap the heads and sentences, but I didn't see a second permute statement to swap them back in the returned values tensor before reshaping.
Thanks for clear explanation. Can you have basic code repository for vision transformer where input and output both are image?
Best 🎉
Thank you 🙏:)
Am I the only one to notice that vsauce refrence 02:28
Pls start LLMs
Just wondering how much you earn given that you know so much about AI and that AI is super hot right now?
Where is the github link?
My bad. I added this in now in the video description
@@CodeEmporium thanks a lot man
First comment
Love it :)
Come ooon dude, we're coding Transformers in 2024, in.... 720p? My eyes hurt :C
It’s a bit funny that you speak perfect English with absolutely no accent, but you pronounce “matrix” wrong. It should be pronounced “may tricks”, not “mah tricks”.
I love my mah-trices
@@CodeEmporium, LOL
at 37:42, shouldn't we permute the values to (1, 4, 8, 64) first and then reshape to (1,4,8*64) to accurately concatenate multiple heads for corresponding sentences?
Thanks!
Thanks!