Perfect! That is one of the best explanations I have ever seen about transformers and the attention mechanism. Very intuitive and at the same time technically rich. God bless you!
Please keep up with this amazing teaching work! Your way of decrypting complex methods in small blocks, while constantly repeating where every parameter comes from really makes it so easy to follow and understand! :D
why is query on the decoder side and key and values are on encoder side? It seems intuitively, but the paper didn't mention it. Did you figure it out by yourself or from somewhere? thx
The attention mechanism is essentially a weighted average. In his previous lecture, he listed two methods to determine the weights of hidden states. The first method is used in Bahdanau et al. (2015)'s paper. The second method as he explained in this video is used in Vaswani et al. (2017)'s paper. Which paper are you referring to?
@@danielmejia8530 every word? It is just a lie. Subtitles auto generation mechanism also have a hard time to parse some words. I just told truth, may be in a direct fashion, and author if he'd appreciate listeners could apply a bit of effort to make his work even more useful for others. Please, go and learn to be honest.
@@vyacheslavvorobyov22 Yes, every word. Hard to believe? He is just using very restriced technical vocab, this is not poetry. Chill and try a little bit harder without crying on TH-cam . And stop calling your absurd biased opinion a ground truth, stop being that cocky and respect others. Lastly, if a machine does better at listening english, please, go for some extra lessons.
@@danielmejia8530 It is you who are started to cry here like an offended girl. And it is exactly you who think his bullshit is a ground truth. It is a right thing when people express his opinion on a TH-cam, because this is a how content could evolve and become better. Feedback is basis for quality. If author decided to be public and share its content here I believe he is ready to critics. But you, saying that "it is all ok, because I understand him!", and exactly at this point you are trying to be a ground truth indicator. I.e. if Mr. Daniel understand everything then this is enough to claim this content is good for all. So, please, put you absolutism right in your ... you know.
@@vyacheslavvorobyov22 I won't argue with a complete asshole who thinks it's right to write 'offended girl' as an insult. You should learn to respect people, creators of content, women. Oh, perhaps you came from a retrograde country? That could be a nice explanation...
I love the fact that you repeat concepts throughout the video. It really hammers the information into the head.
The only resource from which I could understand this topic.
This is the best video series I found on Transformers and Attention on internet
Perfect! That is one of the best explanations I have ever seen about transformers and the attention mechanism. Very intuitive and at the same time technically rich. God bless you!
Please keep up with this amazing teaching work! Your way of decrypting complex methods in small blocks, while constantly repeating where every parameter comes from really makes it so easy to follow and understand! :D
The best lecture about transformers that I've seen 🙏🏻🙏🏻🙏🏻🙏🏻🙏🏻
this is the best video series for transformer. Please continue
THIS is how the teaching should be done
i never seen such clear explanation for complex concept...bravo Wang
Thank you for your excellent lesson, very easy to understand. 5 stars for you.
I really like your approach. very through and clearly and sequentially explained.
Loved it and enjoyed a lot. Thanks. Very good job
Thank you for bringing to us such awesome videos. Best explanations!
Good illustration and explanation. Well done!
This is a great explanation! Thanks so much!!!
Love the videos. Consider dark backgrounds for ease of long-term viewing.
Excellent presentation!
Great Explanation, Thanks!
Amazing video.
Great channel
thank you for sharing; very helpful🙂
Such a good video!
marvelous videos. thanks!
why is query on the decoder side and key and values are on encoder side? It seems intuitively, but the paper didn't mention it. Did you figure it out by yourself or from somewhere? thx
The attention mechanism is essentially a weighted average. In his previous lecture, he listed two methods to determine the weights of hidden states. The first method is used in Bahdanau et al. (2015)'s paper. The second method as he explained in this video is used in Vaswani et al. (2017)'s paper. Which paper are you referring to?
thank you for this amazing video !
can you show coding of it in tensorflow
That's good work. Could have been better if you had explained intuition behind attention and self-attention.
thanks, this is a great lesson!
great explanation thanks
You're amazing
it helped me :D
Author, please provide subtitles, you have a horrible accent.
As a non-native speaker, I understand every word. By the way, that was terribly offensive. Go learn some manners, please.
@@danielmejia8530 every word? It is just a lie. Subtitles auto generation mechanism also have a hard time to parse some words. I just told truth, may be in a direct fashion, and author if he'd appreciate listeners could apply a bit of effort to make his work even more useful for others. Please, go and learn to be honest.
@@vyacheslavvorobyov22 Yes, every word. Hard to believe? He is just using very restriced technical vocab, this is not poetry. Chill and try a little bit harder without crying on TH-cam . And stop calling your absurd biased opinion a ground truth, stop being that cocky and respect others. Lastly, if a machine does better at listening english, please, go for some extra lessons.
@@danielmejia8530 It is you who are started to cry here like an offended girl. And it is exactly you who think his bullshit is a ground truth. It is a right thing when people express his opinion on a TH-cam, because this is a how content could evolve and become better. Feedback is basis for quality. If author decided to be public and share its content here I believe he is ready to critics. But you, saying that "it is all ok, because I understand him!", and exactly at this point you are trying to be a ground truth indicator. I.e. if Mr. Daniel understand everything then this is enough to claim this content is good for all. So, please, put you absolutism right in your ... you know.
@@vyacheslavvorobyov22 I won't argue with a complete asshole who thinks it's right to write 'offended girl' as an insult. You should learn to respect people, creators of content, women. Oh, perhaps you came from a retrograde country? That could be a nice explanation...