EM Algorithm : Data Science Concepts

แชร์
ฝัง
  • เผยแพร่เมื่อ 28 ก.ย. 2024

ความคิดเห็น • 139

  • @anuraganand9675
    @anuraganand9675 2 ปีที่แล้ว +70

    The world needs to see this. Thanks Ritvik, I honestly have utmost respect and love for the amount of hard work you put in your videos. Cheers :)

    • @e555t66
      @e555t66 ปีที่แล้ว +2

      It would take a lot of time to develop these intuitions on your own.

  • @rachelhobbs6189
    @rachelhobbs6189 ปีที่แล้ว +16

    Your channel and way of teaching is so amazing!! Very inviting, inclusive, and friendly. Thank you so much for such good vibes 💕

  • @nnyabavictoria6332
    @nnyabavictoria6332 2 ปีที่แล้ว +5

    Thank you Ritvik for simplifying EM algorithm like this. This is the best video I have seen so far.

  • @adrian-mu3jr
    @adrian-mu3jr 2 ปีที่แล้ว +5

    That's really great way to look at EM. I'm an engineering graduate but new to ML and the workup explanation before dropping into the maths is excellent. thanks

  • @andrashorvath2411
    @andrashorvath2411 ปีที่แล้ว +2

    Awesome explanation. I'd like to extend yours with my intuition regarding the E-Step: the first term p(x|m0) shows the probability of x happening for the chosen m0, and the second term LogLikelihood shows the probability of x happening for the computed m, and we want to maximize both. Because we want a choice with high probability from every aspect. That's why we multiply them together. Because the multiplication can weight between them. If one of them is small then the result will be small. It can be high only if both are high.

    • @ritvikmath
      @ritvikmath  ปีที่แล้ว +1

      thanks for the additional inputs!

  • @vantongerent
    @vantongerent 2 ปีที่แล้ว +5

    YES! I have quiz on this NEXT WEEK!

  • @alizarean5080
    @alizarean5080 8 หลายเดือนก่อน +2

    I have an exam tomorrow and this video was the thing I needed. I can't thank you enough dude.

  • @aaroojyashin3790
    @aaroojyashin3790 8 หลายเดือนก่อน +30

    This could be by far the best explanation I have seen for EM algorithm. The way you have connected the intuitive way to mathematical explanation, is so so commendable!!!! Thank you so much for your efforts

    • @ritvikmath
      @ritvikmath  8 หลายเดือนก่อน

      Glad it was helpful!

    • @chainetravail2439
      @chainetravail2439 หลายเดือนก่อน

      @@ritvikmath I comfirm, thank you for helping lost students

  • @michalistsatsaronis8728
    @michalistsatsaronis8728 2 ปีที่แล้ว +10

    Thanks!You explained such a complicated subject so clearly!!!!

  • @rishi71095
    @rishi71095 ปีที่แล้ว +2

    It would take me two more lives to be able to explain it this well to someone, kudos! Great job buddy!

  • @christophb9616
    @christophb9616 2 ปีที่แล้ว +5

    Thanks for the very clear explanation! A follow up video on how the EM algorithm can be used in gaussian mixture models or bayesian networks would be awesome!

  • @aniketsakpal4969
    @aniketsakpal4969 ปีที่แล้ว +2

    Incredible explanation! Was trying to understand the intuition behind EM for a long time! Thanks for the video! Keep Going!!

  • @peterhopkinson3040
    @peterhopkinson3040 ปีที่แล้ว +1

    Your videos are unreal, simple explanations of complex problems its insane.

  • @navyadewan8680
    @navyadewan8680 5 หลายเดือนก่อน

    your understanding and explanation of such a complicated concept is impeccable

  • @aidanheffernan652
    @aidanheffernan652 ปีที่แล้ว +1

    It's 4am and I saw this video and had to watch....... really great explanation bro.....your a natural teacher.....thanks for this......subscribed

  • @MN-zs8lc
    @MN-zs8lc 4 หลายเดือนก่อน

    Although there is more for fully understanding, I was able to gain the concept because of your video!

  • @TanmayGhonge
    @TanmayGhonge ปีที่แล้ว +1

    Broke down the most complicated algorithm in the simplest terms. Wow!

  • @ehsanmon
    @ehsanmon 2 ปีที่แล้ว +4

    Thanks so much for this great and explanation! I would definitely be interested in the proof. It will be great if you could do a video on Gaussian mixture models as well and how it is solved using the EM algorithm.

  • @sakshamsingh2005
    @sakshamsingh2005 4 หลายเดือนก่อน +1

    By far the best explanation, amazing.

  • @abhishekganapure6456
    @abhishekganapure6456 4 หลายเดือนก่อน

    The only explanation you need for understanding EM algorithm, proper chad explanation!

  • @mohammed2250
    @mohammed2250 2 ปีที่แล้ว

    Yes please go on with the prove, that will be an interesting topic. Though I went on Andrew's ng video couple of times, but I couldn't understand it better than here!! You're a rock star in simplifying complex concepts!!

  • @gergerger53
    @gergerger53 2 ปีที่แล้ว +1

    I had a jolt of excitement when I saw you had decided to do a video on this topic. It's something I've had to revisit time and time again, always understanding the intuition, but always getting lost in the formulas. Your post did a great job at helping to explain the intuition. I did struggle a bit with your non-conventional likelihood notation, though. That did throw me off a little bit but I understand why you had to have it that way and quickly adjusted. The care you took in explaining why there is mu and mu0 just shows why you are a fantastic teacher.

  • @MegaNightdude
    @MegaNightdude 2 ปีที่แล้ว +3

    Great video !

  • @ericwilson8665
    @ericwilson8665 ปีที่แล้ว

    Absolutely fantastic. I agree w/ other comments... The DS world needs to see this. Thank you.

    • @ritvikmath
      @ritvikmath  ปีที่แล้ว

      Glad you enjoyed it!

  • @sahilagarwal1871
    @sahilagarwal1871 2 ปีที่แล้ว +2

    Would love to see a proof video! Keep up the great work!

  • @awangsuryawan7320
    @awangsuryawan7320 2 ปีที่แล้ว +3

    Can't wait to see the proof

  • @nuamaaniqbal6373
    @nuamaaniqbal6373 2 ปีที่แล้ว

    i thank GOD i found your channel. A big thanks to youtube and to you!!

  • @hannahnelson4569
    @hannahnelson4569 หลายเดือนก่อน

    Very cool! Thank you for teaching!

  • @juliachu4685
    @juliachu4685 ปีที่แล้ว

    THANK YOU. You're literally saving my ML undergraduate course

  • @tb1131
    @tb1131 ปีที่แล้ว

    Your explanations are soooo clear! really appreciate the effort you put into your videos. Thank youu!!

  • @Martyr022
    @Martyr022 2 ปีที่แล้ว +5

    I'm interested in the proof!

  • @bradleyadjileye1202
    @bradleyadjileye1202 6 หลายเดือนก่อน +1

    Amazing, thank you for that !

    • @ritvikmath
      @ritvikmath  6 หลายเดือนก่อน

      Glad you liked it!

  • @eceserin
    @eceserin 2 ปีที่แล้ว +2

    Example with python coming anytime?

  • @evgenyivanko9680
    @evgenyivanko9680 8 หลายเดือนก่อน

    Ritvik, you are doing a great job, thanks

  • @JanSchmid-d4k
    @JanSchmid-d4k ปีที่แล้ว

    Thank you for all the work you put in your videos to make life's like mine easier. Cheers man!

  • @dannyzilberg3101
    @dannyzilberg3101 6 หลายเดือนก่อน

    Thank you so much for these videos!
    One question: how do you estimate and maximize the integral in practice? That was the elephant for me...

  • @vijayrajan5792
    @vijayrajan5792 ปีที่แล้ว

    Very compelling ... Brilliant

  • @moravskyvrabec
    @moravskyvrabec ปีที่แล้ว

    Wonderful. Definitely helps my understanding. When I find time I want to see what you're doing w/ the stock predictions. If I remember lectures from business school, you should not be able "generate alpha" unless you possess information the market does not. In this case you could say you've found some new idea that has real predictive value, but either they will a) already have found this and put much more compute + their proximity to the actual place where the trades happen towards getting the answer first and beating you to the trades or b) didn't know it before but will immediately steal it and then see a) haha. But hey, I'll still watch to see what you've got going on.

  • @sharmilakarumuri6050
    @sharmilakarumuri6050 2 ปีที่แล้ว +4

    Can you do the proof too please

  • @nikosp3903
    @nikosp3903 4 หลายเดือนก่อน

    Amazing explanation!

  • @sgklee2664
    @sgklee2664 2 ปีที่แล้ว

    This explanation is amazing in order to get the concept

  • @srinivasbringu67
    @srinivasbringu67 ปีที่แล้ว

    Lovely, that's very intuitive. Thank you so much.

  • @QuocHuyPham-s4n
    @QuocHuyPham-s4n 7 หลายเดือนก่อน

    Excellent explanation!!

  • @dougcree6486
    @dougcree6486 ปีที่แล้ว

    A worked example of the final process would be invaluable.

  • @pallavibothra9671
    @pallavibothra9671 2 ปีที่แล้ว +3

    Hi Ritvik, thank you very much for awesome videos. Could you please make some videos on SQL?

    • @ritvikmath
      @ritvikmath  2 ปีที่แล้ว +2

      thanks! and please check out my full SQL playlist here:
      th-cam.com/play/PLvcbYUQ5t0UFAZGthysGOAtZqLl3otZ-k.html

    • @pallavibothra9671
      @pallavibothra9671 2 ปีที่แล้ว

      @@ritvikmath Awesome! Thanks a lot.. Could you please add sql with window function to the playlist, if possible?

  • @taxtr4535
    @taxtr4535 หลายเดือนก่อน

    You are a gem

  • @biswajit-k
    @biswajit-k ปีที่แล้ว

    Got this crystal clear. Thanks a lot!

  • @nitishyamsukha
    @nitishyamsukha 10 วันที่ผ่านมา

    very good explanation!

    • @ritvikmath
      @ritvikmath  10 วันที่ผ่านมา

      Glad you think so!

  • @sairaamvenkatraman5998
    @sairaamvenkatraman5998 8 หลายเดือนก่อน

    Coild you please do the derivation or intuition for EM for clustering? I observe that it is described in many textbooks, but not in such a cool way. 😅

  • @jalaluddinmohammad262
    @jalaluddinmohammad262 2 ปีที่แล้ว +2

    Please make a proof video

  • @Thefare1234
    @Thefare1234 9 หลายเดือนก่อน

    Great explanation. However, the way you have written it, there is no difference between the likelihood function and the probability function. I think for clarity you should swap x,1,2 and \mu. Also you should use ; instead of | so that the likelihood function is not confused with conditional probability.

  • @tungbme
    @tungbme 2 ปีที่แล้ว

    Thank you so much for your explanation, helps me a lot

  • @sneggo
    @sneggo ปีที่แล้ว

    Great video! Thank you so much

    • @ritvikmath
      @ritvikmath  ปีที่แล้ว

      Glad it was helpful!

  • @Andres186000
    @Andres186000 2 ปีที่แล้ว +3

    Nice to see the theorem guaranteeing convergence for sequences that are increasing and bounded being used to prove this. I do have a more pragmatic question which is how somebody would go about finding the argmax in the M step. Would gradient descent be used on the expectation of log-likelihood function (I would imagine in this case the expectation of log-likelihood would have to be convex for this to work) to find the argmax?

    • @Michael-vs1mw
      @Michael-vs1mw 2 ปีที่แล้ว +2

      Yep, you can use any optimization method. For Gaussian mixture models there are explicit formulas for the M step which are obtained in the usual way by setting the gradient of the expected log-likelihood to zero.

  • @michaelrainer7487
    @michaelrainer7487 ปีที่แล้ว

    On step 2, what does the dx do at the end of that equation?

  • @bbcailiang7
    @bbcailiang7 8 หลายเดือนก่อน

    Excellent explanations!

  • @Tyokok
    @Tyokok 2 ปีที่แล้ว

    Thanks for the great lecture. One question if I may: 2:20, why you put best guess 1 here instead of a random draw from your known distribution?

  • @skanoun
    @skanoun 3 หลายเดือนก่อน

    Great teacher❤

  • @nipa161991
    @nipa161991 ปีที่แล้ว

    Thanks for the video! What was not clear to me is whether we calculate all E(LL|M) for all Ms in which we can calculate the argman in step 3?

  • @kevinshao9148
    @kevinshao9148 ปีที่แล้ว

    Thanks for the great video! One question: if you have (1+2+x)/3 = x , then you can have close form solution, why you still need numerical approach?

  • @grjesus9979
    @grjesus9979 ปีที่แล้ว

    What if x is high dimensional? How would the integral change?

  • @neelabhchoudhary2063
    @neelabhchoudhary2063 6 หลายเดือนก่อน

    oh my god. this was so helpful

    • @ritvikmath
      @ritvikmath  6 หลายเดือนก่อน

      Awesome!

  • @mohammadrezanargesi2439
    @mohammadrezanargesi2439 ปีที่แล้ว

    Something does not seem right to me in the E-step.
    I think the likelihood should be written given the latent variable, which is "x" in this case. But you have written it in given mu...
    I'm confused.
    Also I don't understand how to solve the M-step.
    When i write it down in this case i cannot update x at all🤦🤦🤦
    I only update mu 🤦🤦
    I'm completely confused

  • @lordscourge-jp8ch
    @lordscourge-jp8ch 11 หลายเดือนก่อน

    Thank you so much BRO

  • @shivambindal4809
    @shivambindal4809 2 ปีที่แล้ว

    The expression for Expectation seems similar to Bayesian theory where we have prior belief (P(x|u)) and likelihood and we are multiplying both to get posterior. Is this the same concept?

  • @apresunreve
    @apresunreve ปีที่แล้ว

    amazing, thanks for such a clear explanation :)

  • @EW-mb1ih
    @EW-mb1ih ปีที่แล้ว

    Is the EM algorithm the best algorithm to use in some specific problem (compared for example to the gradient descent algorithm)?

  • @kulbhushansingh1101
    @kulbhushansingh1101 ปีที่แล้ว

    great man, ultra great

  • @akshiwakoti7851
    @akshiwakoti7851 2 ปีที่แล้ว

    Great videos. Got it in one go! Could you do Gaussian Mixture Models? Thanks.

  • @ehsanmon
    @ehsanmon 2 ปีที่แล้ว

    How would the problem change if we didn't know the variance either?

  • @AhmedMohamed-sp4mm
    @AhmedMohamed-sp4mm ปีที่แล้ว

    Thank you so much for these amazing vids.
    Would you kindly provide any MATLAB codes that illustrate these concepts?

  • @oliviernussli3799
    @oliviernussli3799 2 ปีที่แล้ว

    really great explanation! thank you :-)

  • @fowadkhan7605
    @fowadkhan7605 2 ปีที่แล้ว

    Sir. I only know basic statistics. From where should I start watching your videos. Is there any order to them? The concepts you are mentioning I am not familiar with them.

  • @n.m.c.5851
    @n.m.c.5851 2 ปีที่แล้ว

    thank you !!!

  • @underlecht
    @underlecht ปีที่แล้ว

    Interesting way of looking at EM problem. Thank you.

  • @taotaotan5671
    @taotaotan5671 2 ปีที่แล้ว

    Wouldn't integration usually more difficult to do than derivative?

    • @ritvikmath
      @ritvikmath  2 ปีที่แล้ว +2

      Great question! I should have made it more clear that the EM algo is often used in the case of *discrete* latent variables which would make your integral a much more manageable sum.

    • @taotaotan5671
      @taotaotan5671 2 ปีที่แล้ว

      @@ritvikmath Thanks!!

  • @bharathgopalakrishnan3739
    @bharathgopalakrishnan3739 ปีที่แล้ว

    please do the proof video

  • @dhairya931
    @dhairya931 ปีที่แล้ว

    Please provide R code for modelling the dependence between trial and suceess ( herons case by J Zhu) which include em algorithm of beta binomial Poisson mixture model. Please help me

  • @BIBHUTIBHUSANPRADHAN-e3o
    @BIBHUTIBHUSANPRADHAN-e3o ปีที่แล้ว

    How x1=1?...It is again the same value 0!

  • @Fredfetish
    @Fredfetish 2 ปีที่แล้ว

    Can you explain EM algorithm in terms of compositional data please?

  • @simeonvince8013
    @simeonvince8013 ปีที่แล้ว +5

    Thank you for the high-quality contents that you have produced over the past few years. Most of the time, it really did help me get the intuition and understanding of what was going on with the theoretical concepts I was seeing in my courses.
    Once again, thank you !

  • @tecnom7133
    @tecnom7133 5 หลายเดือนก่อน +2

    The best Thanks man

  • @sharmilakarumuri6050
    @sharmilakarumuri6050 2 ปีที่แล้ว +3

    Very well explained

  • @H1K8T95
    @H1K8T95 2 ปีที่แล้ว +2

    Ngl my favorite rapper-turned-algorithm

  • @yulinliu850
    @yulinliu850 2 ปีที่แล้ว +3

    Excellent. Thank you so much! 👍

  • @Darkev77
    @Darkev77 2 ปีที่แล้ว +6

    I’m very confused as of why not just maximize the log-likelihood of all the current observed data given some mu?

    • @michaurbanski5961
      @michaurbanski5961 2 ปีที่แล้ว

      I think it applies not only for estimation of mu, but any arbitrary parameter. Then it would not be as simple as taking average of all observed data. I could be wrong though :P

    • @ritvikmath
      @ritvikmath  2 ปีที่แล้ว +5

      Love this question, thanks for asking. Indeed with this toy example the EM algorithm is overkill and it was mostly meant for instructive purposes. Of course, when we use things for instructive purposes we can miss the more interesting applications. Thinking about this, an interesting variation would be what if you have the data [1,2,x,y] drawn from a N(0,sigma) where now it’s the standard deviation sigma as well as two missing values you want to predict. This is interesting because it’s important to consider the values of the non missing data *and* the potential values of the missing data which are consistent with some estimate for sigma (since standard deviation is inherently a measure between data points)

  • @nawfalguefrachi2071
    @nawfalguefrachi2071 ปีที่แล้ว +1

    thank you ritvik the best videos are in this channel.
    Very intresting way of teaching thank you from TUNISIA

  • @louighi91
    @louighi91 2 ปีที่แล้ว +1

    Holy, i can't believe how good this video was :) thank you so much

  • @Sushanta-Das
    @Sushanta-Das 5 หลายเดือนก่อน

    Sir , I know, In E - step we estimate unknown x , but you are calculating Likelyhood . how are these connected ?

  • @albertonieto2557
    @albertonieto2557 2 ปีที่แล้ว +1

    Brilliant explanation. I especially appreciate you first providing the intuition of the method in the verbal explanation of the E and M steps. I struggled with the seeing the math first in other lectures until seeing your video. Thanks for posting this.

  • @ricardogomes9528
    @ricardogomes9528 ปีที่แล้ว

    What if our distribution is not gaussian, but let’s say poisson, can we trick this algorithm?

  • @bassoonatic777
    @bassoonatic777 2 ปีที่แล้ว +1

    Much better explanation than what I normally see. I would also be interested in seeing you go through the proof.

  • @MohitKumar-vh8ht
    @MohitKumar-vh8ht ปีที่แล้ว

    🥺🥺 Unfortunately I didn't get anything...

  • @Mrfrikha1
    @Mrfrikha1 ปีที่แล้ว

    why x = -1 and not equal 0 im kind confused for his first guess on the first question

  • @dadimanoj9051
    @dadimanoj9051 ปีที่แล้ว

    Thankyou for explaining very clearly

  • @chaurasiaansh6059
    @chaurasiaansh6059 หลายเดือนก่อน

    this is the best lecture for em algo

  • @isciaberger5017
    @isciaberger5017 ปีที่แล้ว

    this helped so much, thank you a lot!!

  • @MalTimeTV
    @MalTimeTV ปีที่แล้ว

    From darkness into the light!

  • @sumedhamandal8062
    @sumedhamandal8062 ปีที่แล้ว

    Damn

  • @yaochung-chen
    @yaochung-chen ปีที่แล้ว

    Really nice explaination! Thank you!

    • @ritvikmath
      @ritvikmath  ปีที่แล้ว

      Glad it was helpful!