What are Maximum Likelihood (ML) and Maximum a posteriori (MAP)? ("Best explanation on YouTube")

แชร์
ฝัง
  • เผยแพร่เมื่อ 5 ก.ย. 2024
  • Explains Maximum Likelihood (ML) and Maximum a posteriori (MAP) estimation/detection using a Gaussian measurement/sampling example.
    * If you would like to support me to make these videos, you can join the Channel Membership, by hitting the "Join" button below the video, and making a contribution to support the cost of a coffee a month. It would be very much appreciated.
    Check out my search for signals in everyday life, by following my social media feeds:
    Facebook: www.facebook.c...
    Instagram: / iainexplains
    Related videos: (see: www.iaincolling...)
    • What is a Probability Density Function (pdf)? • What is a Probability ...
    • What is a Multivariate Probability Density Function (PDF)? • What is a Multivariate...
    • What is a Random Variable? • What is a Random Varia...
    • What is a Gaussian Distribution? • What is a Gaussian Dis...
    • What is Gaussian Noise? • What is Gaussian Noise?
    For a full list of Videos and accompanying Summary Sheets, see the associated website: www.iaincolling...
    .

ความคิดเห็น • 152

  • @ayushkumarrai1117
    @ayushkumarrai1117 3 ปีที่แล้ว +39

    You're the professor I wished I had in my college! Thankyou!!

  • @arvindp551
    @arvindp551 2 ปีที่แล้ว +2

    What you just fabulously explained in 15 lines, takes 4+ blackboards to many Indian teachers to explain even less than that. Thank you so much for sharing your knowledge here.

  • @Peolorios
    @Peolorios 3 ปีที่แล้ว +11

    I found your videos at the right moment, they cover a lot of the basics of my 1st semester master courses. Thank you. A nice topic you could cover that comes up a lot in detection and estimation is the Cramer-Lao Lower Bound

    • @iain_explains
      @iain_explains  3 ปีที่แล้ว +2

      Thanks for your comment. Glad the videos are helpful. And thanks for the C-R suggestion, I'll add it to my "to do" list.

  • @burakkurt3027
    @burakkurt3027 3 ปีที่แล้ว +2

    You explain the concept not only very concise way but also in saving paper. I appreciate you for both the topic and the saved paper.

    • @iain_explains
      @iain_explains  3 ปีที่แล้ว +3

      Thanks. I hadn't realised how environmentally responsible I was being. 😀 I think it really helps to fit everything onto a single sheet of paper so that the whole explanation is visible all at once, so the viewer can easily refer back at any point.

    • @burakkurt3027
      @burakkurt3027 3 ปีที่แล้ว

      @@iain_explains This is very logical :D

  • @junchiliu7350
    @junchiliu7350 2 ปีที่แล้ว +3

    The best explanation on ML and MAP! I finally understood them. Thank you!

    • @iain_explains
      @iain_explains  2 ปีที่แล้ว

      I'm so glad the video helped, and that you liked the explanation.

  • @rahulrustagi7029
    @rahulrustagi7029 7 หลายเดือนก่อน +1

    Amazing. Make a series of Probabilistic ML Models!

  • @maclee3841
    @maclee3841 2 ปีที่แล้ว +1

    Thanks a lot! One of the most simplest explanations on TH-cam

  • @aidankennedy6973
    @aidankennedy6973 3 ปีที่แล้ว +7

    This is a fantastic video that answered so many questions I had while working through my academic coursework. Thank you so much for uploading!

  • @NeilDeshpande313
    @NeilDeshpande313 3 ปีที่แล้ว +2

    I finally get the difference between the two! Thank you!

  • @SouthEsVideos
    @SouthEsVideos 3 ปีที่แล้ว +3

    I'm currently learning about autoencoders and it's based on this topic! very helpful and intuitive. Thank you!

  • @NoahElRhandour
    @NoahElRhandour 3 ปีที่แล้ว

    holy what a clear explanation. it ended my 2 day struggle of not getting it in 18 minutes!!!! thank you

  • @Amine1z
    @Amine1z 3 ปีที่แล้ว +1

    Dear prof, you're the best

    • @iain_explains
      @iain_explains  3 ปีที่แล้ว

      Thanks. Glad you like the videos.

  • @sayantankundu973
    @sayantankundu973 ปีที่แล้ว

    Definitely "Best explanation on TH-cam" !! ❤ Thanks a lot Sir.

    • @iain_explains
      @iain_explains  ปีที่แล้ว

      Thanks. I'm glad you think so. And I'm glad it was helpful.

  • @nikkoNine
    @nikkoNine ปีที่แล้ว

    Very intuitive explanation! 🙏

  • @heroszeros
    @heroszeros ปีที่แล้ว

    Very good explanation with right amount of details and relevant examples. Thanks a million.

  • @usmanshabbir4902
    @usmanshabbir4902 2 ปีที่แล้ว

    very precisely explained.

  • @shahabansari5201
    @shahabansari5201 3 ปีที่แล้ว +1

    Wow! Amazing way of explaining these complex ideas.

  • @zhipenglin
    @zhipenglin หลายเดือนก่อน +1

    excellent video!!

  • @jizong234
    @jizong234 2 ปีที่แล้ว

    This is the best explanation in the world, thank you !

    • @iain_explains
      @iain_explains  2 ปีที่แล้ว +1

      I'm so glad to hear that you liked the video.

  • @waynerooney4674
    @waynerooney4674 ปีที่แล้ว +1

    ty

  • @JoaoVitorBRgomes
    @JoaoVitorBRgomes 3 ปีที่แล้ว +1

    Beautiful explanation. Very helpful.

  • @peterkimemiah9669
    @peterkimemiah9669 2 ปีที่แล้ว

    Thanks for your super simple explanation. I now understand how to apply it.

  • @ivanrodionov9724
    @ivanrodionov9724 ปีที่แล้ว +1

    Outstanding video! You sir have saved the day, again!

    • @iain_explains
      @iain_explains  ปีที่แล้ว +1

      I'm so glad the video helped. It's great to read these comments, and know that my videos are making a difference for people. Thanks.

  • @ashokmahipathi8133
    @ashokmahipathi8133 ปีที่แล้ว

    Thank you so much Sir Iain. You made my day. Great explanation regrading MAP and ML. Hats off Iain

  • @prietessutube
    @prietessutube 2 ปีที่แล้ว

    Thanks so much for this video, explained it much better with my lecturer!!!

  • @abderrahmanetayebcherif7325
    @abderrahmanetayebcherif7325 2 ปีที่แล้ว

    best explanation of the ML and MAP on youtube
    thank you

  • @AleeEnt863
    @AleeEnt863 ปีที่แล้ว

    Great explanation...

  • @KCODivyanshuKatyan
    @KCODivyanshuKatyan ปีที่แล้ว +1

    best in the game 🙌🙌

  • @alexanderpaskal1387
    @alexanderpaskal1387 2 ปีที่แล้ว

    Liked and subbed, very clear and accessible explanation of a concept that made no sense to me as it was presented in my class

  • @passionatechristianmaker
    @passionatechristianmaker 6 หลายเดือนก่อน

    great explanation!

    • @iain_explains
      @iain_explains  6 หลายเดือนก่อน

      Glad it was helpful!

  • @Dutta1605
    @Dutta1605 9 หลายเดือนก่อน

    i beleive the title of the video is genuinely true.

    • @iain_explains
      @iain_explains  9 หลายเดือนก่อน

      Thanks. It was a comment someone else had made about the video, so it's good to know that you also agree.

  • @robsenponte3308
    @robsenponte3308 ปีที่แล้ว +1

    Nice

  • @Balance-fl1zc
    @Balance-fl1zc 8 หลายเดือนก่อน

    Beautiful explanation sir, thank you!

  • @keremokyay6725
    @keremokyay6725 ปีที่แล้ว

    Amazing, this was so clear to understand. Thank you very much!!!

  • @saraho5338
    @saraho5338 2 ปีที่แล้ว

    Thank you for the great video

  • @Yohan87845
    @Yohan87845 2 ปีที่แล้ว

    Thank you very much.

  • @aqeelal-shakhouri7572
    @aqeelal-shakhouri7572 2 ปีที่แล้ว

    Thank you. you explained it clearly, just what I was looking for.

  • @ashleyhart1907
    @ashleyhart1907 11 หลายเดือนก่อน

    Incredibly helpful. Thank you!

    • @iain_explains
      @iain_explains  11 หลายเดือนก่อน

      You're very welcome!

  • @AbCd-fo6ys
    @AbCd-fo6ys 2 ปีที่แล้ว

    What a clear explanation!
    Thank you so much.

  • @mehmetfatihayten8269
    @mehmetfatihayten8269 ปีที่แล้ว

    Decent video! Thanks.

  • @user-pq5jy3yo8v
    @user-pq5jy3yo8v 26 วันที่ผ่านมา

    Thank you!!!

  • @rezasoleimani6636
    @rezasoleimani6636 ปีที่แล้ว

    The explanation is great. The only problem is using pen and paper instead of something more comfortable. The page is too small for this amount of writing.

    • @iain_explains
      @iain_explains  ปีที่แล้ว +1

      On the other hand, having everything on the one page means you don't need to scroll back and forth through the video to see the links to earlier parts, and I can simply point to the earlier parts while explaining how they link to the later parts (as I'm doing in the thumbnail image). Perhaps it doesn't work so well on small screens ...

  • @gastonparra2971
    @gastonparra2971 ปีที่แล้ว

    Gran explicación.. Gracias por subir el video.

    • @iain_explains
      @iain_explains  ปีที่แล้ว

      My pleasure. I'm glad you liked it.

  • @mangomilkshakelol
    @mangomilkshakelol 8 หลายเดือนก่อน

    Thank you so much! That's clear. One question: for MAP, what's f_X(x)?

    • @iain_explains
      @iain_explains  7 หลายเดือนก่อน +1

      It's the probability density function for the variable X.

  • @loreychu7646
    @loreychu7646 3 ปีที่แล้ว

    Good explanation of a lot of concepts in wireless communication. I'm watching your video for the preparation of QE. Hope I can pass!

    • @iain_explains
      @iain_explains  3 ปีที่แล้ว

      Glad it was helpful! Good luck!

  • @adonis7419
    @adonis7419 3 ปีที่แล้ว +1

    Very helpful

  • @user-ns9ze8xf5z
    @user-ns9ze8xf5z 3 ปีที่แล้ว

    Really big thanks for your video!!
    May you take another video for explaining different pathloss models, such as Okumura-Hata or various COST model in wireless channel?

    • @iain_explains
      @iain_explains  3 ปีที่แล้ว +2

      Good suggestions thanks. I've added them to my "to do" list.

  • @gofaonef.mogobe1306
    @gofaonef.mogobe1306 3 ปีที่แล้ว +1

    Hi..very helpful video. Kindly assist me understand how I can factor in the concept of consistency of MLE with respect to the graph illustrations?
    Particularly, I've learned that as n gets large, mean turns to zero as MLE becomes an even more consistent estimate.

  • @partheshvikassoni8601
    @partheshvikassoni8601 2 ปีที่แล้ว

    This was really informative! Thanks.

  • @cartercheng5846
    @cartercheng5846 2 ปีที่แล้ว

    Very helpful video. I have a question there. The MAP is explained as MLE weighed by the probability of the parameter x, and the parameter follows a certain distribution. If X is a continuous random variable, what is the mathematical meaning for f_y(y|x)f_x(x)?

    • @iain_explains
      @iain_explains  2 ปีที่แล้ว

      It equals the joint pdf f_{X,Y}(x,y)

  • @ibrahimshikdaher7551
    @ibrahimshikdaher7551 3 ปีที่แล้ว +1

    Thank you

  • @ks.4494
    @ks.4494 ปีที่แล้ว

    Thanks for the Video, is there any reference ( book, ...) for that, particulary for numerical solution?

    • @iain_explains
      @iain_explains  ปีที่แล้ว

      I'm not sure if this is what you're looking for exactly (eg. I'm not sure it has the numerical examples you might be looking for), but I like this book: H. Vincent Poor, “An Introduction to Signal Detection and Estimation”

  • @ronaldmtenga8093
    @ronaldmtenga8093 ปีที่แล้ว

    are the differenct x values you are checking for maximum likelihood each a possible input signal or are we searching on a bit by bit basis?

    • @iain_explains
      @iain_explains  ปีที่แล้ว

      They are a possible realisation of the random variable X. If X represents binary data, then it would be "searching on a bit by bit basis", but it X represented higher order modulation then it would be on a "symbol" basis.

  • @AK-yf4dp
    @AK-yf4dp 2 ปีที่แล้ว

    Thank you so much!!! very helpful video

  • @yuxinzhang4228
    @yuxinzhang4228 2 ปีที่แล้ว

    It's very helpful thanks sooooo much

    • @iain_explains
      @iain_explains  2 ปีที่แล้ว

      You're welcome! I'm glad it helped.

  • @BlueSkyGoldSun
    @BlueSkyGoldSun ปีที่แล้ว

    Can this be applied in marketing?

  • @tuongnguyen9391
    @tuongnguyen9391 3 ปีที่แล้ว

    Do you ever have to do a rehearsal beforehand ? I see the explanation is quite smooth.

    • @iain_explains
      @iain_explains  3 ปีที่แล้ว +1

      Thanks, I put quite a bit of thought into how to explain things in the best way.

    • @tuongnguyen9391
      @tuongnguyen9391 3 ปีที่แล้ว

      Do you follow a systematic procedure to construct the explanation process. If so I really hope that you could share this procedure :). Although everything is short I find that the information is delivered clearly with many subtle points and detail carefully summarized. Thank you for your inspiring lecture.

  • @javiergonzalezarmas8250
    @javiergonzalezarmas8250 ปีที่แล้ว

    Amazing

  • @joaquinvanloon7341
    @joaquinvanloon7341 2 ปีที่แล้ว

    Amazing video, thanks!

  • @AlexColeAldems
    @AlexColeAldems 2 ปีที่แล้ว

    Does demodulating using ML require channel state information? (i.e. an estimation of the AWGN noise variance)

    • @iain_explains
      @iain_explains  2 ปีที่แล้ว +1

      Yes. Good point. It's almost never mentioned. It's not too hard to get an estimate of the receiver noise - by taking measurements when nothing is being sent (of course you need to be able to work out when nothing is being sent!) It's harder to estimate other parameters, such as channel gain. And there's lots of things that are done to make that possible. See eg: "Channel Estimation for Mobile Communications" th-cam.com/video/ZsLh01nlRzY/w-d-xo.html

  • @tainacoimbra1940
    @tainacoimbra1940 3 ปีที่แล้ว

    Excelent explanation! Thank you very much :)

  • @davebordenkircher8131
    @davebordenkircher8131 ปีที่แล้ว

    I see now that the MAP estimator is like a weighted version of the ML estimator, where the weights come from prior knowledge of the measurement target. The different conditional distributions fy(y|xi) are “pushed up” or “pushed down” based on the value of the corresponding fx(xi). Of course, provided that all fx(xi) are equiprobable, the MAP estimator reduces to the ML estimator which we commonly see in optimal communications system analysis.
    I have a question for you, why is it that the equiprobable symbol scheme is considered most optimal? I am inclined to assume that it is because it yields the highest entropy. Also, I would like to know how it is that we ensure equiprobable signaling?
    Thank you

    • @iain_explains
      @iain_explains  ปีที่แล้ว +1

      Excellent question. Yes, when a random source is compressed to its minimal representation (using an entropy achieving codebook) it results in a binary sequence that has equally likely ones and zeros. This video provides more insights: "What is Entropy? and its relation to Compression" th-cam.com/video/FlaJPxP8sd8/w-d-xo.html

  • @_Sam_-zh7sw
    @_Sam_-zh7sw 2 ปีที่แล้ว

    may be i am missing some pre-requisite knowledge because i am confused a little bit. have we inverted the graph of the function here? f(x) is plotted horizontally and x is plotted vertically. But how can there be a different distribution function of ax1,ax2...ax(n) if there is just one input and output?

    • @iain_explains
      @iain_explains  2 ปีที่แล้ว

      Um no. x is not plotted vertically. f_Y(y|x) is plotted vertically. This is the density of the random variable Y, given a specific value of the random variable X. This is a different function for each different realisation (value) of X (ie. x_1, x_2, ...).

  • @saipavan7703
    @saipavan7703 2 ปีที่แล้ว

    Hi Iain. Loved your explanation. I wanted to ask a question about MLE. In the plots of x1,x2,xn, When each x1/x2 give a single value for the function, Why does plot exist for x1 when the function takes a single value for x1. Thank You

    • @iain_explains
      @iain_explains  2 ปีที่แล้ว

      Sorry, I'm not sure what you're asking.

    • @rishabhsrivastava1241
      @rishabhsrivastava1241 2 ปีที่แล้ว

      I think I got what you are saying and there seems to be some gap in your understanding. Let me try to fill that although Ian mentioned it in this video.
      What you are saying is that for a given value of x, there is only one single value of y through it's distribution f(y/x) but that is not true. Actually, there are SEVERAL different distributions of y depending on the SEVERAL values of x's. So, when Ian says that for a given x, the distribution's center shifts, it is actually a new distribution centered around that given x value. Then comes the concept of a single value from these distributions, now that is y(bar), this is an observation of all the f(y/x) pdf value among all the distributions of y's for those SEVERAL x's. That is the single value that you are thinking of.
      Hope I was able to answer your question to some degree. :)

    • @mrr7115
      @mrr7115 2 ปีที่แล้ว

      The single value is a result of having measured/fixed y (the y bar hat equation). The plot is for all y (ie it’s a function of y not of x1). x1 is just a guess of the true parameter of the Gaussian distribution (proportional to mean). The horizontal axis (independent variable) is y. Also the function, which is a Gaussian, takes more than just a*x1, it also takes in the variance from the noise. To prove to yourself that the function takes y, look at the form of the Gaussian equation, see the y in there?

  • @akinsolaidris9292
    @akinsolaidris9292 5 วันที่ผ่านมา

    Is AwGN the same as normally distributed err or bias?

    • @iain_explains
      @iain_explains  3 วันที่ผ่านมา

      This video should help: "What is White Gaussian Noise (WGN)?" th-cam.com/video/QfUQMzHfbxs/w-d-xo.html

  • @kavineshs15
    @kavineshs15 2 หลายเดือนก่อน

    I can't understand why the bell curve is shifting for every value of x.

  • @NvkChaitanya777
    @NvkChaitanya777 2 ปีที่แล้ว

    Can you please make a video on softmax regression?

    • @iain_explains
      @iain_explains  2 ปีที่แล้ว

      Thanks for the suggestion, but I'm not familiar with it, sorry. I'll have to give it some thought.

  • @ravindratomar9916
    @ravindratomar9916 2 ปีที่แล้ว

    sir, how to estimate channel in case of correlated rayleigh fading channel. for example y1=hx_1 + hx_2 +n_1, y2=hx_1 + hx_2 +n_2.

    • @ravindratomar9916
      @ravindratomar9916 2 ปีที่แล้ว

      n_1 and n_2 are white gaussian noise with different variances.

  • @robertpollock8617
    @robertpollock8617 10 หลายเดือนก่อน

    So it is L(x|y)? - we want to maximize the likelihood of x given the data values y? . So we are in a sense trying to say that we have high likelihood that this data observed could come from or be predicted by this model of x? Where the probability is P(y|x). Maybe you are saying that and I am not picking up on this. I think you might be but I might not be understanding your notation.

    • @iain_explains
      @iain_explains  10 หลายเดือนก่อน

      Yes, that's right.

  • @kalyansunkara9464
    @kalyansunkara9464 2 ปีที่แล้ว

    Sir which text book should we follow for detection and estimation theory?

    • @iain_explains
      @iain_explains  2 ปีที่แล้ว +1

      I like the book: H. Vincent Poor, "An Introduction to Signal Detection and Estimation", Springer.

  • @analysislearning9179
    @analysislearning9179 3 ปีที่แล้ว

    Great video, I have a question, if the variable is its self distributed with Nakagami distribution. Then how can we compute the MLE and MAP?

    • @iain_explains
      @iain_explains  3 ปีที่แล้ว +1

      The term f_X(x) is the density function for the random variable of interest. So, if it is Nakagami distributed, then f_X(x) equals the formula for the Nakagami p.d.f. which you can find in this video: th-cam.com/video/ztpNbE-Vpaw/w-d-xo.html

    • @analysislearning9179
      @analysislearning9179 3 ปีที่แล้ว +1

      @@iain_explains Thankyou very much

  • @donpkchannel7203
    @donpkchannel7203 2 ปีที่แล้ว

    When your typing the screens becomes blurry because paper is moving. Please stabilize the paper.

  •  2 ปีที่แล้ว +1

    MAP starts at 10:35

  • @amitpalkar934
    @amitpalkar934 3 ปีที่แล้ว

    Can we have a video sometime on mmse and irc receivers ?
    Regards,
    Amit

    • @iain_explains
      @iain_explains  3 ปีที่แล้ว

      Thanks for the suggestion. I've added them to my "to do" list. I'll see what I can do (it's starting to become a long list).

  • @gamingandmusic9217
    @gamingandmusic9217 3 ปีที่แล้ว

    at a particular point in the density function, the probability is zero right? I'm a little confused.

    • @iain_explains
      @iain_explains  3 ปีที่แล้ว

      I'm not exactly sure what you're asking. The density function is a "density" (as the name indicates). This means you need to integrate it over some range of values, in order to find the probability. The probability of any _exact_ value is zero (since the base has zero width, for a single _exact_ value). See: "What is a Probability Density Function (pdf)?" th-cam.com/video/jUFbY5u-DMs/w-d-xo.html

    • @gamingandmusic9217
      @gamingandmusic9217 3 ปีที่แล้ว

      @@iain_explains oh sorry, im wrong. Thank you so much sir.

  • @human_agi
    @human_agi ปีที่แล้ว

    if x is vector ?

    • @iain_explains
      @iain_explains  ปีที่แล้ว

      Yes, it all follows though for vectors.

  • @skatingscotland64
    @skatingscotland64 2 ปีที่แล้ว

    Iain you have offered me shelter in a howling wind, thank you - I can leave the library and go home now xo love from rory

    • @iain_explains
      @iain_explains  2 ปีที่แล้ว

      That's great. I'm so glad you found the video helpful. Hope you mange to stay out of the wind.

  • @egorkrapovnitskiy997
    @egorkrapovnitskiy997 ปีที่แล้ว

    didn't get the idea

  • @tomaguci5102
    @tomaguci5102 ปีที่แล้ว

    I think this video could be improved by providing a concrete example, also it's really mathy without much intuitive explanation

  • @MohammedFarag81
    @MohammedFarag81 11 หลายเดือนก่อน

    Great, thank you

    • @iain_explains
      @iain_explains  11 หลายเดือนก่อน

      Glad you liked it!