A Short Introduction to Entropy, Cross-Entropy and KL-Divergence

แชร์
ฝัง
  • เผยแพร่เมื่อ 4 ก.พ. 2018
  • Entropy, Cross-Entropy and KL-Divergence are often used in Machine Learning, in particular for training classifiers. In this short video, you will understand where they come from and why we use them in ML.
    Paper:
    - "A mathematical theory of communication", Claude E. Shannon, 1948, pubman.mpdl.mpg.de/pubman/item...
    Errata:
    * At 5:05, the sign is reversed on the second line, it should read: "Entropy = -0.35 log2(0.35) - ... - 0.01 log2(0.01) = 2.23 bits"
    * At 8:43, the sum of predicted probabilities should always add up to 100%. Just pretend that I wrote, say, 23% instead of 30% for the Dog probability and everything's fine.
    The painting on the first slide is by Annie Clavel, a great French artist currently living in Los Angeles. The painting is reproduced with her kind authorization. Please visit her website: www.annieclavel.com/.
  • วิทยาศาสตร์และเทคโนโลยี

ความคิดเห็น • 464

  • @revimfadli4666
    @revimfadli4666 4 ปีที่แล้ว +438

    This feels like a 1.5-hour course conveyed in just 11 minutes, i wonder how much entropy it has :)

    • @grjesus9979
      @grjesus9979 3 ปีที่แล้ว +2

      hahaha

    • @anuraggorkar5595
      @anuraggorkar5595 3 ปีที่แล้ว +1

      Underrated Comment

    • @klam77
      @klam77 3 ปีที่แล้ว +3

      ahhh....too clever. the comment has distracted my entropy from the video. Negative marks for you!

    • @Darkev77
      @Darkev77 3 ปีที่แล้ว

      @@klam77 Could you elaborate on his joke please?

    • @ashrafg4668
      @ashrafg4668 2 ปีที่แล้ว +4

      @@Darkev77 The idea here is that most other resources (videos, blogs) take a very long time (and more importantly say a lot of things) to convey the ideas that this video did in a short time (and with just the essential ideas). This video, thus, has low entropy (vs most other resources that have much higher entropy).

  • @jennyread9464
    @jennyread9464 6 ปีที่แล้ว +555

    Fantastic video, incredibly clear. Definitely going to subscribe!
    I do have one suggestion. I think some people might struggle a little bit around 2m22s where you introduce the idea that if P(sun)=0.75 and P(rain)=0.25, then a forecast of rain reduces your uncertainty by a factor of 4. I think it's a little hard to see why at first. Sure, initially P(rain)=0.25 while after the forecast P(rain)=1, so it sounds reasonable that that would be a factor of 4. But your viewers might wonder why you can’t equally compute this as, initially P(sun)=0.75 while after the forecast P(sun)=0. That would give a factor of 0!
    You could talk people through this a little more, e.g. say imagine the day is divided into 4 equally likely outcomes, 3 sunny and 1 rainy. Before, you were uncertain about which of the 4 options would happen but after a forecast of rain you know for sure it is the 1 rainy option - that’s a reduction by a factor of 4. However after a forecast of sun, you only know it is one of the 3 sunny options, so your uncertainty has gone down from 4 options to 3 - that’s a reduction by 4/3.

    • @AurelienGeron
      @AurelienGeron  6 ปีที่แล้ว +57

      Thanks Jenny! You're right, I went a bit too fast on this point, and I really like the way you explain it. :)

    • @god-son-love
      @god-son-love 6 ปีที่แล้ว +1

      Shouldn't one use information gain to check the extent of reduction ? IG = (-1log2(1) - 0log2(0) ) - (-(3/4)log2(4/3)-(1/4)log2(1/4)) = 0.01881437472 bit

    • @dlisetteb
      @dlisetteb 5 ปีที่แล้ว +3

      thank youuuuuuuuuuuuuuuuu

    • @rameshmaddali6208
      @rameshmaddali6208 5 ปีที่แล้ว +19

      Actually I understand the concept from your comment than the video itself :) thanks a lot

    • @maheshwaranumapathy4678
      @maheshwaranumapathy4678 5 ปีที่แล้ว +8

      awesome, great insight i did struggle to get it at first place. Checked out the comments and bam! Thanks :)

  • @ArxivInsights
    @ArxivInsights 6 ปีที่แล้ว +257

    As a Machine Learning practitioner & TH-cam vlogger, I find these videos incredibly valuable! If you want to freshen up on those so-often-needed theoretical concepts, your videos are much more efficient and clear than reading through several blogposts/papers. Thank you very much!!

    • @AurelienGeron
      @AurelienGeron  6 ปีที่แล้ว +19

      Thanks! I just checkout out your channel and subscribed. :)

    • @pyeleon5036
      @pyeleon5036 6 ปีที่แล้ว +2

      I like your video too! Especially the VAE one

    • @fiddlepants5947
      @fiddlepants5947 5 ปีที่แล้ว +5

      Arxiv, it was actually your video on VAE's that encouraged me to check out this video for KL-Divergence. Keep up the good work, both of you.

    • @grjesus9979
      @grjesus9979 4 ปีที่แล้ว

      thank you, at first i messed up trying to understand but now reading your comment i understamd it. Thank you! 😊

  • @xintongbian
    @xintongbian 6 ปีที่แล้ว +41

    I've been googling KL Divergence for some time now without understanding anything... your video conveys that concept effortlessly. beautiful explanation

  • @011azr
    @011azr 6 ปีที่แล้ว +4

    Sir, you have a talent to explain stuff in a crystal clear manner. You just make something that is usually explained by a huge sum of math equations to be something so simple like this. Great job, please continue on making more TH-cam videos!

  • @sushilkhadka8069
    @sushilkhadka8069 20 วันที่ผ่านมา

    Wow best explaination ever, I found this while I was in college. I just come here once a year just to refresh my intution.

  • @jdm89s13
    @jdm89s13 5 ปีที่แล้ว +1

    This 11-ish minute presentation so clearly and concisely explained what I had a hard time understanding from a one hour lecture in school. Excellent video!

  • @hassanmatout741
    @hassanmatout741 6 ปีที่แล้ว +2

    This channel will sky rocket. no doubt. Thank you so much! Clear, visualized and well explained at a perfect pace! Everything is high quality! Keep it up sir!

  • @glockenspiel_
    @glockenspiel_ 3 ปีที่แล้ว +2

    Thank you, very well explained! I decided to get into machine learning in this hard quarantine period but I didn't have many expectations placed on me. Thanks to your clear and friendly explanations in your book I am learning, improving and, not least, enjoying a lot. So thank you so much!

  • @jackfan1008
    @jackfan1008 5 ปีที่แล้ว +1

    This explanation is absolutely fantastic. Clear, concise and comprehensive. Thank you for the video.

  • @chenranxu6941
    @chenranxu6941 3 ปีที่แล้ว +1

    Wow! It's just incredible to convey so much information while still keeping everything simple & well-explained, and within 10 min.

  • @s.r8081
    @s.r8081 3 ปีที่แล้ว +1

    Fantastic! This short video really explains the concept of entropy, cross-entropy, and KL-Divergence clearly, even if you know nothing about them before.
    Thank you for the clear explaination!

  • @user-gd7br4uk5m
    @user-gd7br4uk5m 6 ปีที่แล้ว +7

    I come to find Entorpy, but I received Entorpy, Cross-Enropy and KL-divergence. You are so generous!

  • @aa-xn5hc
    @aa-xn5hc 6 ปีที่แล้ว +40

    you are a genius in creating clarity

  • @LC-lj5kd
    @LC-lj5kd 6 ปีที่แล้ว +2

    ur tutorial is always invincible. quite explicit with great examples. Thanks for ur work

  • @agarwaengrc
    @agarwaengrc ปีที่แล้ว +1

    Haven't seen a better, clearer explanation of entropy and KL-Divergence, ever, and I've studied information theory before, in 2 courses and 3 books. Phenomenal, this should be made the standard intro for these concepts, in all university courses.

  • @michaelzumpano7318
    @michaelzumpano7318 ปีที่แล้ว +3

    Wow! This was the perfect mix of motivated examples and math utility. I watched this video twice. The second time I wrote it all out. 3 full pages! It’s amazing that you could present all these examples and the core information in ten minutes without it feeling rushed. You’re a great teacher. I’d love to see you do a series on Taleb’s books - Fat Tails and Anti-Fragility.

  • @metaprog46and2
    @metaprog46and2 3 ปีที่แล้ว +2

    Phenomenal explanation of a seemingly esoteric concept into one that's simple & easy-to-understand. Great choice of examples too. Very information-dense yet super accessible for most people (I'd imagine).

  • @homo-sapiens-dubium
    @homo-sapiens-dubium 6 ปีที่แล้ว +6

    Im so happy that I found your channel and youre making such great videos!
    As a computer Science student, truly understanding those concepts is the essence of learning them. Videos like this one help enormously by giving a simple mounting point for understanding using intuition!
    Great work! Keep it going!
    P.S. Your book is also great, can recommend it for everyone really trying to understand ML, not just applying it...

  • @bingeltube
    @bingeltube 6 ปีที่แล้ว +2

    Very recommendable! Finally, I found someone who could explain these concepts of entropy, cross entropy in very intuitive ways

  • @romanmarakulin7448
    @romanmarakulin7448 5 ปีที่แล้ว +1

    Thank you so much! Not only it helped me understand KL-Divergence, but also it is helpful to remember the formula. From now I will place signs in right places. Keep it up!

  • @mohamadnachabe1
    @mohamadnachabe1 4 ปีที่แล้ว +1

    This was the best intuitive explanation of entropy and cross entropy I've seen. Thanks!

  • @Dinunzilicious
    @Dinunzilicious 3 ปีที่แล้ว

    Incredibly video, easily one of the top three I've ever stumbled across in terms of concise educational value. Also love the book, great for anyone wanting this level of clarity on a wide range of ml topics.
    Not sure if this will help anyone else, but I was having trouble understanding why we choose 1/p as the "uncertainty reduction factor," and not, say 1-p or some other metric. What helped me gain an intuition for this was realizing 1/p is the number of bits we would need to encode a uniform distribution if every event had the probability p. So the information, -log(p), is how many bits that event would be "worth" were it part of a uniform distribution. This uniform distribution is also the maximum entropy distribution that event could possibly come from given its probability...though you can't reference entropy without first explaining information.

  • @JakeMiller2020
    @JakeMiller2020 4 ปีที่แล้ว

    I always seem to come back to watch this video every 3-6 months, when I forget what KL Divergence is conceptually. It's a great video.

  • @salman3112
    @salman3112 6 ปีที่แล้ว +2

    Your channel has become one of my favorite channels. Your explanation of CapsNet and now this is just amazing. I am going to get your book too. Thanks a lot. :)

  • @summary7428
    @summary7428 2 ปีที่แล้ว

    this is by far the best and most concise explanation on the fundamental concepts of information theory we need for machine learning..

  • @sagnikbhattacharya1202
    @sagnikbhattacharya1202 6 ปีที่แล้ว +3

    You make the toughest concepts seem super easy! I love your videos!!!

  • @matthewwilson2688
    @matthewwilson2688 6 ปีที่แล้ว +2

    This is the best explanation of entropy and KL I have found. Thanks

  • @meerkatj9363
    @meerkatj9363 6 ปีที่แล้ว +1

    I've seen all your videos now. You've taught me a lot of things and this was some good moments. Can't wait for more. Thanks so much

  • @shiliseifeddine5140
    @shiliseifeddine5140 4 ปีที่แล้ว

    this is by far the best description of those 3 terms , can't be thankful enough

  • @voraciousdownloader
    @voraciousdownloader 4 ปีที่แล้ว +1

    Really the best explanation of KL divergence I have seen so far !! Thank you.

  • @Rafayak
    @Rafayak 5 ปีที่แล้ว +24

    Finally, someone who understands, and doesn't just regurgitate the wikipedia page :) Thanks alot!

  • @frankcastle3288
    @frankcastle3288 3 ปีที่แล้ว

    I have been using cross-entropy for classification for years and I just understood it. Thanks Aurélien!

  • @swapanjain892
    @swapanjain892 6 ปีที่แล้ว +1

    You have no idea how much this video has helped me.Thanks for making such quality content and keep creating more.

  • @yb801
    @yb801 10 หลายเดือนก่อน +1

    Thank you , I have always confused about these three concepts, you make these concepts really clear for me.

  • @colletteloueva13
    @colletteloueva13 10 หลายเดือนก่อน +1

    One of the most beautiful videos I've watched and understood a concept :')

  • @SagarYadavIndia
    @SagarYadavIndia ปีที่แล้ว

    Beautiful short video, explaining the concept that is usually a 2 hour explanation in about 10 minutes.

  • @sunrisetofu
    @sunrisetofu 6 ปีที่แล้ว

    Fantastic series of videos, looking forward to every new one! Thanks for taking the time out to make these

  • @khaledelsayed762
    @khaledelsayed762 2 ปีที่แล้ว

    Very elegant indicating how cognizant the presenter is.

  • @AladinxGonca
    @AladinxGonca 2 หลายเดือนก่อน +1

    You are the most talented tutor I've ever seen

  • @GreenCowsGames
    @GreenCowsGames ปีที่แล้ว

    I am new to information theory and computer science in general, and this is the best explanation I could find about these topics by far!

  • @ykkim77
    @ykkim77 4 ปีที่แล้ว +1

    This is the best explanation of the topics that I have ever seen. Thanks!

  • @ashutoshnirala5965
    @ashutoshnirala5965 4 ปีที่แล้ว

    Thankyou for such a wonderful and to the point video. Now I know: Entropy, Cross Entropy, KL Divergence and also why cross entropy is such a good choice as loss function.

  • @jamesjenkins9480
    @jamesjenkins9480 2 ปีที่แล้ว +1

    I've learned about this before, but this is the best explanation I've come across. And was a helpful review, since it's been a while since I used this. Well done.

  • @EinSteinLiT
    @EinSteinLiT 6 ปีที่แล้ว

    very clear and well-structured explanation. Your book is great, too!Thank you very much!

  • @GuilhermeKodama
    @GuilhermeKodama 5 ปีที่แล้ว +1

    the best explanation I ever had about the topic. It was really insightful.

  • @se123acabaron
    @se123acabaron 5 ปีที่แล้ว

    Fantastic video! It made me understand and get together many "loose" concepts. Thank you very much for this contribution!

  • @tensorfreitas
    @tensorfreitas 6 ปีที่แล้ว +1

    Your book and your videos are incredible. Thank you !

  • @akshiwakoti7851
    @akshiwakoti7851 4 ปีที่แล้ว

    Hats off! One of the best teachers ever! This definitely helped me better understand it both mathematically and intuitively just in a single watch. Thanks for reducing my 'learning entropy'. My KL divergence on this topic is near zero now. ;)

  • @paulmendoza9736
    @paulmendoza9736 ปีที่แล้ว

    I want to like this video 1000 times. To the point, no BS, clear, understandable.

  • @davidbeauchemin3046
    @davidbeauchemin3046 6 ปีที่แล้ว

    Awesome video, you made the concept of entropy so much clearer.

  • @MrFurano
    @MrFurano 6 ปีที่แล้ว

    To-the-point and intuitive explanation and examples! Thank you very much! Salute to you!

  • @laura_uzcategui
    @laura_uzcategui 4 ปีที่แล้ว

    Really good explanation, the visuals were also great for understanding! Thanks Aurelien.

  • @sagarsaxena7202
    @sagarsaxena7202 5 ปีที่แล้ว +1

    Great work in the explanation. I have been pretty confused with this concept and the implication of Information theory with ML. This video does the trick in clarifying the concepts while providing a sync between information theory and ML usage. Thanks much for the video.

  • @maryamzarabian4617
    @maryamzarabian4617 2 ปีที่แล้ว

    thank you for useful video , and also really thanks for your book . You express very difficult concepts of machine learning like a piece of cake .

  • @zoeye720
    @zoeye720 5 ปีที่แล้ว +2

    This video explains the concepts so well! Thank you!

  • @billmo6824
    @billmo6824 2 ปีที่แล้ว

    Really, I definitely cannot come up with an alternative way to explain this concept more concisely.

  • @srikumarsastry7473
    @srikumarsastry7473 6 ปีที่แล้ว +1

    So much clear explanation! Need more of them!

  • @areejabdu3125
    @areejabdu3125 5 ปีที่แล้ว

    this explanation really helps the learner in understanding such vague scientific concepts, thanx for the clear explanation !!

  • @pyeleon5036
    @pyeleon5036 6 ปีที่แล้ว +1

    It's so good to watch your video! Thank you so much!

  • @thegamersschool9978
    @thegamersschool9978 2 ปีที่แล้ว

    I am reading your book! and oh man oh what a book!!! first I thought how the book and video has exactly same example for explanation until I saw the book of yours on the later part of the video, and realized it's you it's so great to listen to you after reading you!!

  • @julioreyram
    @julioreyram 3 ปีที่แล้ว

    I'm amazed by this video, you are a gifted teacher.

  • @leastactionlab2819
    @leastactionlab2819 4 ปีที่แล้ว +1

    Great video to learn interpretations of the concept of cross-entropy.

  • @chinmaym92
    @chinmaym92 5 ปีที่แล้ว +2

    I rarely comment on videos, but this video is so good. I just couldn't resist. Thank you so much for the video. :)

  • @YYchen713
    @YYchen713 2 ปีที่แล้ว

    Fantastic video! Now all the dots are connected! I have used loss function for NN machine learning, but not knowing the math behind it! This is so enlightening!

  • @elvisng1977
    @elvisng1977 2 ปีที่แล้ว

    This video is so clear and so well explained, just like his book!

  • @fberron
    @fberron 2 ปีที่แล้ว

    Finally I understood Shannon's theory of information. Thank you
    Aurélien

  • @0ManishSharma0
    @0ManishSharma0 5 ปีที่แล้ว

    Wow, this is great. Thank you for the detailed and clear explanation.

  • @Dr.Roxirock
    @Dr.Roxirock 10 หลายเดือนก่อน +1

    I really enjoyed the way you are explaining it. It's so inspiring watching and learning difficult concepts from the author of such an incredible book in the ML realm. I wish you could teach via video other concepts as well.
    Cheers,
    Roxi

  • @shuodata
    @shuodata 4 ปีที่แล้ว

    Best Entropy and Cross-Entropy explanation I have ever seen

  • @unleasedflow8532
    @unleasedflow8532 3 ปีที่แล้ว

    Nicely conveyed what is to be learned about the topic. I think I absorbed all the way. Best tutorial, keep dropping video like this.

  • @ramonolivier57
    @ramonolivier57 3 ปีที่แล้ว +1

    Excellent explanation and discussion. Thank you very much!!

  • @rohitkumarz
    @rohitkumarz 5 ปีที่แล้ว +1

    i saw many video and then i stumbled on your video.. so much informative and very well articulated thankyou once more.. will check out your book.

  • @andrewtwigg
    @andrewtwigg 2 ปีที่แล้ว

    Thanks for the explanation, very clear and complements your excellent book

  • @deteodskopje
    @deteodskopje 4 ปีที่แล้ว

    Very nice. Really short yet clearly grasping the point of these concepts. Subscribed.
    I was really excited when I found this chanel. I mean the book Hands On Machine Learning is maybe the best book you can find these days

  • @misnik1986
    @misnik1986 3 ปีที่แล้ว

    Thank you so much Monsieur Geron pour cette explication simple et limpide

  • @ramensusho
    @ramensusho หลายเดือนก่อน

    The no. of bits I received is way higher than I expected !!
    Nice video

  • @michaelding5970
    @michaelding5970 4 ปีที่แล้ว

    The best explanation I've seen on this topic.

  • @AbhishekSingh-og7kf
    @AbhishekSingh-og7kf 3 ปีที่แล้ว

    Every concept are very clear... Thanks a lot!!

  • @frkake
    @frkake 6 ปีที่แล้ว

    Thank you very much!
    Excellent video. I started to read your book.
    I respect you.

  • @VincentKun
    @VincentKun ปีที่แล้ว +1

    Ok, i maybe should pay more attention when reading my books, but when i heard here that CrossEntropy is entropy + KL it made sense, then when i read my notes i wrote something similar, but without even realizing how big it was.

  • @danyalkhaliq915
    @danyalkhaliq915 4 ปีที่แล้ว

    super clear .. never I heard this explanation of Entropy and Cross Entropy !

  • @vinithapalani3385
    @vinithapalani3385 6 ปีที่แล้ว +1

    Will definitely check out your book..your videos are amazing...please keep them coming!!

  • @ekbastu
    @ekbastu 5 ปีที่แล้ว +2

    I came here to learn how to correctly pronounce his name :).
    The content is simply great. Thanks a lot.

  • @will.a.i.3405
    @will.a.i.3405 4 ปีที่แล้ว

    Simply awesome. Thank you for such great explanation!

  • @JoeVaughnFarsight
    @JoeVaughnFarsight ปีที่แล้ว

    Merci Aurélien Géron, c'était une très belle présentation !

  • @CowboyRocksteady
    @CowboyRocksteady ปีที่แล้ว

    i'm loving the slides and explaination. I noticed the name in the corner and thought, oh nice i know that name. then suddenly... It's the author of that huge book i love!

  • @vaishanavshukla5199
    @vaishanavshukla5199 4 ปีที่แล้ว +2

    great understanding
    and very good mentor

  • @MrMijertone
    @MrMijertone 6 ปีที่แล้ว +1

    I had to find a word for how well you explain. Perspicious. Thank you.

    • @AurelienGeron
      @AurelienGeron  6 ปีที่แล้ว

      I just learned a new word, thanks James! :)

  • @anonymous.youtuber
    @anonymous.youtuber ปีที่แล้ว +1

    Magnificent explanation! 👍

  • @tarun9305
    @tarun9305 5 ปีที่แล้ว +1

    Such a great explanation! Thank you.

  • @paulstahlhofen5017
    @paulstahlhofen5017 5 ปีที่แล้ว +1

    Finally got the point of what all this stuff actually means. Thanks a lot! My lecturers could learn from you. Just subscribed

  • @DiogoSanti
    @DiogoSanti 4 ปีที่แล้ว

    Awesome video! Hope you deliver more content here very soon!

  • @WanderlustBites215
    @WanderlustBites215 2 ปีที่แล้ว

    Beautifully explained. Thank you!!

  • @gowthamramesh2443
    @gowthamramesh2443 5 ปีที่แล้ว +11

    Kinda feels like 3Blue1Brown's version of Machine learning Fundamentals. Simply Amazing

    • @AurelienGeron
      @AurelienGeron  5 ปีที่แล้ว +5

      Thanks a lot, I'm a huge fan of 3Blue1Brown! 😊

  • @leoxu9673
    @leoxu9673 2 ปีที่แล้ว

    This is fantastic. Thank you so much for this and your book!

  • @sc0tty319
    @sc0tty319 6 ปีที่แล้ว +1

    Great explaination! Very intuitive examples I love it!! keep it comingg!!

  • @annwang2990
    @annwang2990 4 ปีที่แล้ว

    Very clear and explainable, I bought your book! Thank you!

  • @Vladeeer
    @Vladeeer 6 ปีที่แล้ว

    I have that book, didn't realized you wrote it until now.

  • @aiyifei4732
    @aiyifei4732 3 ปีที่แล้ว

    Thanks, explain is clear. I found it's clean and easy to understand compare with my lecture notes. I don't even think they mentioned the history and derivation/origin