Maximum Likelihood Estimation - THINK PROBABILITY FIRST!

แชร์
ฝัง
  • เผยแพร่เมื่อ 11 ม.ค. 2025

ความคิดเห็น • 43

  • @ArashSadr
    @ArashSadr 2 ปีที่แล้ว +4

    I am just falling in love with the way you break down the problem and go after each part!!

  • @sujathaontheweb3740
    @sujathaontheweb3740 หลายเดือนก่อน

    This channel should already have a million likes and 10 million subscribers!! 🎉🌈💛💡

  • @GoingData
    @GoingData 11 หลายเดือนก่อน +3

    I have been studying stats and maths for years... this guy is INCREDIBLE!, is like khan academy... i can understand.... this is a real blessing to my thirst for knowledge. Thank you !!!

  • @zgbjnnw9306
    @zgbjnnw9306 3 ปีที่แล้ว +6

    I learned both point estimate and MLE in my statistics class,I had no idea they are related somehow! Thanks for this inspiring video!

  • @wish.e.9925
    @wish.e.9925 2 ปีที่แล้ว +4

    I have just watched two videos and want to say that the logic you use to explain things is the best approach for delivering such knowledge. Thank you so much.

  • @shchen16
    @shchen16 2 ปีที่แล้ว +2

    I've just watch two of your videos and honestly it has saved my life.
    Thanks for such great video!

  • @bikinibottom2100
    @bikinibottom2100 2 ปีที่แล้ว +1

    I'm going to look at every video in this channel, I can't wait to see the next ones, specially love the ones about sampling !

  • @mahlatseseneke5894
    @mahlatseseneke5894 2 ปีที่แล้ว +1

    You are heaven sent... May the Most High turn everything you touch into gold... You deserve a Nobel Prize :,)

    • @KapilSachdeva
      @KapilSachdeva  2 ปีที่แล้ว

      🙏 Not sure if I deserve this much appreciation but many thanks for your kindness and best wishes.

    • @mahlatseseneke5894
      @mahlatseseneke5894 2 ปีที่แล้ว

      @@KapilSachdeva will you be doing any more videos for this book?

    • @KapilSachdeva
      @KapilSachdeva  2 ปีที่แล้ว

      Yes.

  • @a_j_lifestyle6515
    @a_j_lifestyle6515 10 หลายเดือนก่อน +1

    Brilliantly done. Connecting the dots and democratizing ML math

  • @ArulManokaran
    @ArulManokaran 9 หลายเดือนก่อน +1

    Kapil, I will be forever greatful to you for this lecture series.

  • @rajibkhan9749
    @rajibkhan9749 2 ปีที่แล้ว +1

    Thanks @Kapil Sachdeva . Its helping a lot

  • @spyrosp.551
    @spyrosp.551 11 หลายเดือนก่อน +1

    Great work sir, thank you

  • @sanjaythorat
    @sanjaythorat 2 ปีที่แล้ว

    Excellent series. It is helping me understand Bayesian Stats/Regression better. Thanks for the series. Keep doing more.
    I have few questions though.
    - At around 14:00, you are multiplying normal distributions together. Shouldn't it be multiplication of probabilities of target values given normal distribution?
    - 20:45 - Wouldn't MSE help us derive uncertainty?

    • @KapilSachdeva
      @KapilSachdeva  2 ปีที่แล้ว

      - in any case we have assume variables to be Gaussian in this example to resulting variable after multiplication will be Gaussian.
      - the tutorial showed the relation between MSE and MLE but by itself it is not quantifying the uncertainty. May be I have misunderstood your question? Follow up if my answer is not clear.

    • @sanjaythorat
      @sanjaythorat 2 ปีที่แล้ว

      ​@@KapilSachdeva Let me rephrase my questions:
      - At around 14:00, you are multiplying normal distributions together. But, I think you should be multiplying the probabilities of the target variables given the parameters of the normal distribution. isn't it? It's probably just the notation that I do not understand completely.
      - 20:45 - I think we should be able to compute uncertainty/variance of the target variables using mean squared error computed with linear regression, we probably do not need Bayesian regression for that as you claimed.

    • @KapilSachdeva
      @KapilSachdeva  2 ปีที่แล้ว +1

      it's the notation and inconsistent vocabulary. its the distribution functions that are getting multiplied (note - in reality you would take log and add them and not multiply)
      You could do that but thinking in probability first is a more natural approach to obtain the information

  • @bodwiser100
    @bodwiser100 ปีที่แล้ว

    Wonderful video! One question -- we did not make use anywhere of the fact that the function f is a polynomial in x (unless I am mistaken, embarrassingly). So it means that we can have any generic set of features on the right hand side instead of only powers of x. Then what was the reason for assuming that f is a polynomial in x?

  • @ztoza7158
    @ztoza7158 3 ปีที่แล้ว +1

    Awesome! Will need some example to understand this better . But great start for me. Thanks fir putting this together

    • @KapilSachdeva
      @KapilSachdeva  3 ปีที่แล้ว

      🙏 The example is same as that of part 1. If you use the notebook from part 1, all you need to do is to compute the precision using the analytical expression shown in this tutorial.

  • @tobe7602
    @tobe7602 2 ปีที่แล้ว +1

    Thanks for this great video. Can you explain me why probability of a point is not Zero in the formula p(t0).p(t1)...p(tn), i always learn p(x) is 0 for a pdf. We must speak lileihood better l(t0).l(t1)...l(tn), who are associate with N(t0,sygma) and N(t1,sygma)...and N(tn,sygma) ?
    You are a very good pedagogue.
    Thanks for you time.
    Antonio

    • @KapilSachdeva
      @KapilSachdeva  2 ปีที่แล้ว +1

      🙏 It is the likelihood. Unfortunately, the symbol used is still p :(

  • @danmathewsrobin5991
    @danmathewsrobin5991 3 ปีที่แล้ว +2

    Great series. Waiting for more of it :) Do you solely depend on the mentioned textbook content for creating the video? Could you mention some other similar resources? I wish somebody would make a series like this around Machine Learning: A Probabilistic Perspective by Kevin Murphy :D. Kudos to the good work around Bishop's book anyways. And may I know how do you make these kind of animations in your videos? Thank you, Kapil. ~dan

    • @KapilSachdeva
      @KapilSachdeva  3 ปีที่แล้ว +2

      Thanks Dan for the appreciation & interest in this series.
      > Do you solely depend on the mentioned textbook content for creating the video?
      Not at all; except for few videos that will be part of this series as I am planning to use the examples from Dr. Bishop's book and hence providing the deserved citation & credit. Also, I believe it would be helpful to read the book after watching the videos in this series or vice versa :). I am aiming to publish few more videos in the series this week.
      > Could you mention some other similar resources?
      My process of learning is quite ad-hoc but if I were to recommend a book on this topic it would be Statistical Rethinking (xcelab.net/rm/statistical-rethinking/) by Dr. McElreath. There are various ports of this book using various toolkits including one done by me using Tensorflow Probability. IMHO, it would be better to read this book before PRML or Kevin Murphy's book.
      In terms of resources, read various papers, thesis, and books as it is important to get different perspectives and explanations of the same subject. More importantly, always seek the "why" .... question & challenge everything ... implement the concepts from scratch and it would do wonders.
      > And may I know how do you make these kind of animations in your videos?
      I use PowerPoint for these animations. Am also quite familiar with manim (by 3blue1brown) now but have not used to create any video so far. I keep coming back to PowerPoint; so far it has worked quite well. I use the "morph" transition quite a lot to animate things.
      Hope this answers some of your questions if not all.

  • @Daydream_Dynamo
    @Daydream_Dynamo 7 หลายเดือนก่อน

    One stupid question here, Why we were interested in finding max joint probability, in the first place?? were there any other way to find w and beta??

  • @Shashank_Shahi1989
    @Shashank_Shahi1989 2 ปีที่แล้ว

    Where is part 1 of this video ? Link please. Which playlist in your channel should be watch first ? Any chronological order or it's just random videos on ML topic ?

    • @KapilSachdeva
      @KapilSachdeva  2 ปีที่แล้ว

      Link to the playlist. You should be able to see it at the channel page as well.
      th-cam.com/play/PLivJwLo9VCUISiuiRsbm5xalMbIwOHOOn.html

  • @ericzhang4486
    @ericzhang4486 3 ปีที่แล้ว

    In this baby step case, since each target has the same sigma/precision, dose it mean that the precision of predictions of the model will always be the same which is beta? In practical problems, we will have different betas for each target, right?

    • @KapilSachdeva
      @KapilSachdeva  3 ปีที่แล้ว

      Yes, it would be better to get the sigma for each target variable.

  • @ssshukla26
    @ssshukla26 3 ปีที่แล้ว

    One question sir, the minimization part, it's derivative of NLL w.r.t w and B and then equate them (both the equations) to 0?

    • @KapilSachdeva
      @KapilSachdeva  3 ปีที่แล้ว

      That is correct.
      here is one good reference. It uses variance instead of precision but you will see the full derivation.
      www.cs.princeton.edu/courses/archive/fall18/cos324/files/mle-regression.pdf