Gradient Boost Part 3 (of 4): Classification

แชร์
ฝัง
  • เผยแพร่เมื่อ 6 ต.ค. 2024

ความคิดเห็น • 543

  • @statquest
    @statquest  4 ปีที่แล้ว +26

    NOTE: Gradient Boost traditionally uses Regression Trees. If you don't already know about Regression Trees, check out the 'Quest: th-cam.com/video/g9c66TUylZ4/w-d-xo.html Also NOTE: In Statistics, Machine Learning and almost all programming languages, the default base for the log function, log(), is log base 'e' and that is what I use here.
    Support StatQuest by buying my book The StatQuest Illustrated Guide to Machine Learning or a Study Guide or Merch!!! statquest.org/statquest-store/

    • @parijatkumar6866
      @parijatkumar6866 3 ปีที่แล้ว +1

      I am a bit confused. The first Log that you took : Log(4/2) - was that to some base other than e? Cause e^(log(x)) = x for log to the base e
      And hence the probability will be simply 2/(1+2) = 2/3 = No of Yes / Total Obs = 4/6 = 2/3
      Pls let me know if this is correct.

    • @statquest
      @statquest  3 ปีที่แล้ว +2

      @@parijatkumar6866 The log is to the base 'e', and yes, e^(log(x)) = x. However, sometimes we don't have x, we just have the log(x), as is illustrated at 9:45. So, rather than use one formula at one point in the video, and another in another part of the video, I believe I can do a better job explaining the concepts if I am consistent.

    • @jonelleyu1895
      @jonelleyu1895 ปีที่แล้ว

      For Gradient Boost for CLASSIFICATION, because we convert the categorical targets(No or Yes) to probabilities(0-1) and the residuals are calculated from the probabilities, when we build a tree, we still use REGRESSION tree, which use sum of squared residuals to split the tree. Is it correct? Thank you.

    • @statquest
      @statquest  ปีที่แล้ว +1

      @@jonelleyu1895 Yes, even for classification, the target variable is continuous (probabilities instead of Yes/No), and thus, we use regression trees.

  • @weiyang2116
    @weiyang2116 3 ปีที่แล้ว +171

    I cannot imagine the amount of time and effort used to create these videos. Thanks!

    • @statquest
      @statquest  3 ปีที่แล้ว +31

      Thank you! Yes, I spent a long time working on these videos.

  • @primozpogacar4521
    @primozpogacar4521 4 ปีที่แล้ว +22

    Love these videos! You deserve a Nobel prize for simplifying machine learning explanations!

    • @statquest
      @statquest  4 ปีที่แล้ว

      Wow, thanks!

  • @sameepshah3835
    @sameepshah3835 3 หลายเดือนก่อน +2

    Thank you so much Josh, I watch 2-3 videos everyday of your machine learning playlist and it just makes my day. Also the fact that you reply to most of the people in the comments section is amazing. Hats off. I only wish the best for you genuinely.

    • @statquest
      @statquest  3 หลายเดือนก่อน +1

      bam!

    • @sameepshah3835
      @sameepshah3835 3 หลายเดือนก่อน +1

      @@statquest Double Bam!
      Bam?

    • @enicay7562
      @enicay7562 19 วันที่ผ่านมา

      @@sameepshah3835 Triple Bam!

  • @OgreKev
    @OgreKev 4 ปีที่แล้ว +54

    I'm enjoying the thorough and simplified explanations as well as the embellishments, but I've had to set the speed to 125% or 150% so my ADD brain can follow along.
    Same enjoyment, but higher bpm (bams per minute)

    • @statquest
      @statquest  4 ปีที่แล้ว +3

      Awesome! :)

  • @jagunaiesec
    @jagunaiesec 4 ปีที่แล้ว +34

    The best explanation I've seen so far. BAM! Catchy style as well ;)

    • @statquest
      @statquest  4 ปีที่แล้ว

      Thank you! :)

    • @arunavsaikia2678
      @arunavsaikia2678 4 ปีที่แล้ว +1

      @@statquest are the individual trees which are trying to predict the residuals regression trees?

    • @statquest
      @statquest  4 ปีที่แล้ว

      @@arunavsaikia2678 Yes, they are regression trees.

  • @dhruvjain4774
    @dhruvjain4774 4 ปีที่แล้ว +10

    you really explain complicated things in very easy and catchy way.
    i like the way you BAM

  • @debsicus
    @debsicus 3 ปีที่แล้ว +52

    This content shouldn’t be free Josh. So amazing Thank You 👏🏽

    • @statquest
      @statquest  3 ปีที่แล้ว +3

      Thank you very much! :)

  • @rameshh3821
    @rameshh3821 2 หลายเดือนก่อน +1

    You have explained the Gradient Boosting Regressor and Classifier very well. Thank you!

    • @statquest
      @statquest  2 หลายเดือนก่อน

      Thank you!

  • @xinjietang953
    @xinjietang953 ปีที่แล้ว +2

    Thanks for all you've done. You know your videos is first-class and precision-promised learning source for me.

    • @statquest
      @statquest  ปีที่แล้ว

      Great to hear!

  • @asdf-dh8ft
    @asdf-dh8ft 3 ปีที่แล้ว +2

    Thank you very much! Your step by step explanation is very helpful. It gives to people with poor abstract thinking like me chance to understand all math of these algorithms.

    • @statquest
      @statquest  3 ปีที่แล้ว +1

      Glad it was helpful!

  • @CodingoKim
    @CodingoKim ปีที่แล้ว +8

    my life has been changed for 3 times. First, when I met Jesus. Second, when I found out my true live. Third, it's you Josh

    • @statquest
      @statquest  ปีที่แล้ว +4

      Triple bam! :)

  • @igormishurov1876
    @igormishurov1876 4 ปีที่แล้ว +7

    Will recommend the channel for everyone study the machine learning :) Thanks a lot, Josh!

    • @statquest
      @statquest  4 ปีที่แล้ว +1

      Thank you! :)

  • @narasimhakamath7429
    @narasimhakamath7429 4 ปีที่แล้ว +4

    I wish I had a teacher like Josh! Josh, you are the best! BAAAM!

    • @statquest
      @statquest  4 ปีที่แล้ว

      Thank you!:)

  • @mayankamble2588
    @mayankamble2588 6 หลายเดือนก่อน +1

    This is amazing. This is the nth time I have come back to this video!

    • @statquest
      @statquest  6 หลายเดือนก่อน

      BAM! :)

  • @cmfrtblynmb02
    @cmfrtblynmb02 3 ปีที่แล้ว +1

    Finally a video that shows the process of gradent boosting. Thanks a lot.

  • @lemauhieu3037
    @lemauhieu3037 2 ปีที่แล้ว +2

    I'm new to ML and these contents are gold. Thank you so much for the effort!

    • @statquest
      @statquest  2 ปีที่แล้ว

      Glad you like them!

  • @Valis67
    @Valis67 3 ปีที่แล้ว +2

    That's an excellent lesson and a unique sense of humor. Thank you a lot for the effort in producing these videos!

    • @statquest
      @statquest  3 ปีที่แล้ว

      Glad you like them!

  • @juliocardenas4485
    @juliocardenas4485 2 ปีที่แล้ว +1

    Yet again. Thank you for making concepts understandable and applicable

  • @soujanyapm9595
    @soujanyapm9595 3 ปีที่แล้ว +1

    Amazing illustration of a complicated concept. This is best explanation. Thank you so much for all your efforts in making us understand the concepts very well !!! Mega BAM !!

    • @statquest
      @statquest  3 ปีที่แล้ว

      Thank you! :)

  • @dankmemer9563
    @dankmemer9563 4 ปีที่แล้ว +2

    Thanks for the video! I’ve been going on a statquest marathon for my job and your videos have been really helpful. Also “they’re eating her...and then they’re going eat me!....OH MY GODDDDDDDDDDDDDDD!!!!!!”

    • @statquest
      @statquest  4 ปีที่แล้ว +1

      AWESOME!!!

  • @amitv.bansal178
    @amitv.bansal178 2 ปีที่แล้ว +1

    Absolutely wonderful. You are are my guru and a true salute to you

  • @tymothylim6550
    @tymothylim6550 3 ปีที่แล้ว +2

    Thank you Josh for another exciting video! It was very helpful, especially with the step-by-step explanations!

    • @statquest
      @statquest  3 ปีที่แล้ว +1

      Hooray! I'm glad you appreciate my technique.

  • @rishabhkumar-qs3jb
    @rishabhkumar-qs3jb 3 ปีที่แล้ว +1

    Fantastic video , I was confused about the gradient boosting, after watching all parts of gb technique from this channel, I understood it very well :)

  • @FabricioVladimirVincesVinces
    @FabricioVladimirVincesVinces 2 หลายเดือนก่อน +1

    It is perfectly understood. Thank you so much!

    • @statquest
      @statquest  2 หลายเดือนก่อน

      Glad it was helpful!

  • @user-gr1qk3gu4j
    @user-gr1qk3gu4j 5 ปีที่แล้ว +1

    Very simple and practical lesson. I did created a worked sample based on this with no problems.
    It might be obvious, but not explained there, that initial mean odd should be more than 1. It might be explained as odd of more rare event should be closer to zero.
    Glad to see this video arrived just at the time I started to interest this topic.
    I guess it will become a "bestseller"

  • @umeshjoshi5059
    @umeshjoshi5059 4 ปีที่แล้ว +2

    Love these videos. Starting to understand the concepts. Thank you Josh.

    • @statquest
      @statquest  4 ปีที่แล้ว

      Thank you! :)

  • @marjaw6913
    @marjaw6913 2 ปีที่แล้ว +1

    Thank you so much for this series, I understand everything thanks to you!

  • @ugwuchiagozie6990
    @ugwuchiagozie6990 หลายเดือนก่อน +1

    God bless you josh
    I really appreciate

    • @statquest
      @statquest  หลายเดือนก่อน

      Thank you!

  • @ΒΑΣΙΛΗΣ_ΛΕΒΕΝΤΑΡΟΣ
    @ΒΑΣΙΛΗΣ_ΛΕΒΕΝΤΑΡΟΣ 18 วันที่ผ่านมา +1

    First of all, I would like to thank you, Dr. Josh, for all these great videos. I would like to ask how important, in your experience, it is to understand the algorithms mathematics, as you analyze them in parts 2 and 4, especially for people who want to work in the analysis of biological data. Thanks a lot again! you really helped me understand many machine learning topics.

    • @statquest
      @statquest  18 วันที่ผ่านมา +1

      One of the reasons I split these videos into "main ideas" and "mathematical details" was I felt that the "main ideas" were more important for most people. The details are interesting, and helpful if you want to build your own tree based method, but not required.

    • @ΒΑΣΙΛΗΣ_ΛΕΒΕΝΤΑΡΟΣ
      @ΒΑΣΙΛΗΣ_ΛΕΒΕΝΤΑΡΟΣ 17 วันที่ผ่านมา +1

      @statquest Thank you for your reply! Also, I would like to thank you again for all this knowledge that you provide. I have never seen a better teaching methodology than yours ! :)

    • @statquest
      @statquest  16 วันที่ผ่านมา

      @@ΒΑΣΙΛΗΣ_ΛΕΒΕΝΤΑΡΟΣ Thank you very much! :)

  • @aweqweqe1
    @aweqweqe1 2 ปีที่แล้ว +1

    Respect and many thanks from Russia, Moscow

  • @rrrprogram8667
    @rrrprogram8667 5 ปีที่แล้ว +2

    I have beeeeennnn waiting for this video..... Awesome job Joshh

  • @HamidNourashraf
    @HamidNourashraf 10 หลายเดือนก่อน +1

    the best video for GBT

    • @statquest
      @statquest  10 หลายเดือนก่อน

      Thanks!

  • @Just-Tom
    @Just-Tom 4 ปีที่แล้ว +3

    I was wrong! All your songs are great!!!
    Quadruple BAM!

  • @SoukainaGhafel
    @SoukainaGhafel 7 หลายเดือนก่อน +1

    thanks alot , ur videos helped me too much, plz keep going

    • @statquest
      @statquest  7 หลายเดือนก่อน

      Thank you!

  • @SergioPolimante
    @SergioPolimante 2 ปีที่แล้ว +1

    man, you videos are just super good, really.

  • @gonzaloferreirovolpi1237
    @gonzaloferreirovolpi1237 5 ปีที่แล้ว +1

    Already waiting for Part 4...thanks as always Josh!

    • @statquest
      @statquest  5 ปีที่แล้ว +1

      I'm super excited about Part 4 and should be out in a week and a half. This week got a little busy with work, but I'm doing the best that I can.

  • @abhilashsharma1992
    @abhilashsharma1992 4 ปีที่แล้ว +5

    Best original song ever in the start!

    • @statquest
      @statquest  4 ปีที่แล้ว +2

      Yes! This is a good one. :)

  • @sidagarwal43
    @sidagarwal43 4 ปีที่แล้ว +1

    Amazing and Simple as always. Thank You

    • @statquest
      @statquest  4 ปีที่แล้ว

      Thank you very much! :)

  • @siyizheng8560
    @siyizheng8560 4 ปีที่แล้ว +2

    All your videos are super amazing!!!!

    • @statquest
      @statquest  4 ปีที่แล้ว

      Thank you! :)

  • @vans4lyf2013
    @vans4lyf2013 3 ปีที่แล้ว +7

    I wish I could give you the money that I pay in tuition to my university. It's ridiculous that people who are paid so much can't make the topic clear and comprehensible like you do. Maybe you should do teaching lessons for these people. Also you should have millions of subscribers!

    • @statquest
      @statquest  3 ปีที่แล้ว

      Thank you very much!

  • @ElderScrolls7
    @ElderScrolls7 4 ปีที่แล้ว +1

    Another great lecture by Josh Starmer.

    • @statquest
      @statquest  4 ปีที่แล้ว +1

      Hooray! :)

    • @ElderScrolls7
      @ElderScrolls7 4 ปีที่แล้ว +1

      @@statquest I actually have a draft paper (not submitted yet) and included you in the acknowledgements if that is ok with you. I will be very happy to send it to you when we have a version out.

    • @statquest
      @statquest  4 ปีที่แล้ว +1

      @@ElderScrolls7 Wow! that's awesome! Yes, please send it to me. You can do that by contacting me first through my website: statquest.org/contact/

    • @ElderScrolls7
      @ElderScrolls7 4 ปีที่แล้ว +1

      @@statquest I will!

  • @ayahmamdouh8445
    @ayahmamdouh8445 2 ปีที่แล้ว +1

    Hi Josh, great video.
    Thank you so much for your great effort.

  • @abissiniaedu6011
    @abissiniaedu6011 ปีที่แล้ว +1

    Your are very helpful, thank you!

  • @timothygorden7689
    @timothygorden7689 ปีที่แล้ว +1

    absolute gold

  • @yulinliu850
    @yulinliu850 5 ปีที่แล้ว +1

    Excellent as always! Thanks Josh!

    • @statquest
      @statquest  5 ปีที่แล้ว

      Thank you! :)

  • @이형석-g9m
    @이형석-g9m 2 ปีที่แล้ว +1

    Great video! Thank you!

  • @vinayakgaikar154
    @vinayakgaikar154 ปีที่แล้ว +1

    nice explanation and easy to understand thanks bro

    • @statquest
      @statquest  ปีที่แล้ว

      You are welcome

  • @patrickyoung5257
    @patrickyoung5257 4 ปีที่แล้ว +5

    You save me from the abstractness of machine learning.

  • @koderr100
    @koderr100 2 ปีที่แล้ว +1

    thanks for videos. best of anything else I did see. Will use this 'pe-pe-po-pi-po" as message alarm on phone)

  • @abyss-kb8qy
    @abyss-kb8qy 4 ปีที่แล้ว +2

    God bless you , thanks you so so so much.

    • @statquest
      @statquest  4 ปีที่แล้ว

      Thank you! :)

  • @suryan5934
    @suryan5934 4 ปีที่แล้ว +4

    Now I want to watch Troll 2

    • @statquest
      @statquest  4 ปีที่แล้ว +1

      :)

    • @AdityaSingh-lf7oe
      @AdityaSingh-lf7oe 4 ปีที่แล้ว +2

      Somewhere around the 15 min mark I made up my mind to search this movie on google

    • @suryan5934
      @suryan5934 4 ปีที่แล้ว

      @@AdityaSingh-lf7oe bam

  • @parthsarthijoshi6301
    @parthsarthijoshi6301 3 ปีที่แล้ว +1

    THIS IS A BAMTABULOUS VIDEO !!!!!!

  • @vijaykumarlokhande1607
    @vijaykumarlokhande1607 2 ปีที่แล้ว

    I salute your hardwork, and mine too

  • @siddharthvm8262
    @siddharthvm8262 2 ปีที่แล้ว +1

    Bloody awesome 🔥

  • @Mars7822
    @Mars7822 2 ปีที่แล้ว +1

    Super Cool to understand and study, Keep Up master..........

  • @nayrnauy249
    @nayrnauy249 2 ปีที่แล้ว +1

    Josh my hero!!!

  • @sid9426
    @sid9426 4 ปีที่แล้ว +1

    Hey Josh,
    I really enjoy your teaching. Please make some videos on XG Boost as well.

    • @statquest
      @statquest  4 ปีที่แล้ว

      XGBoost Part 1, Regression: th-cam.com/video/OtD8wVaFm6E/w-d-xo.html
      Part 2 Classification: th-cam.com/video/8b1JEDvenQU/w-d-xo.html
      Part 3 Details: th-cam.com/video/ZVFeW798-2I/w-d-xo.html
      Part 4, Crazy Cool Optimizations: th-cam.com/video/oRrKeUCEbq8/w-d-xo.html

  • @haitaowu5888
    @haitaowu5888 3 ปีที่แล้ว

    Hi, I have a few questions: 1. How do we know when GBDT algorithms stops( except the M, number of trees) 2. how do I choose value for the M, how do I know this is optimal ?
    Nice work by the way, best explanation I found on the internet.

    • @statquest
      @statquest  3 ปีที่แล้ว

      You can stop when the predictions stop improving very much. You can try different values for M and plot predictions after each tree and see when predictions stop improving.

    • @haitaowu5888
      @haitaowu5888 3 ปีที่แล้ว +1

      @@statquest thank you!

  • @anshvashisht8519
    @anshvashisht8519 ปีที่แล้ว +1

    really liked this intro

  • @jongcheulkim7284
    @jongcheulkim7284 2 ปีที่แล้ว +1

    Thank you so much.

  • @AmelZulji
    @AmelZulji 5 ปีที่แล้ว +1

    First of all thank you for such a great explanations. Great job!
    It would be great if you could make a video about the Seurat package, which very powerful tool for single cell RNA analysis.

  • @anusrivastava5373
    @anusrivastava5373 3 ปีที่แล้ว +1

    Simply Awesome!!!!!!

    • @statquest
      @statquest  3 ปีที่แล้ว

      Thank you! :)

  • @しゅんぷ-x9r
    @しゅんぷ-x9r 4 ปีที่แล้ว +1

    Thank you for good videos!

  • @rrrprogram8667
    @rrrprogram8667 5 ปีที่แล้ว +2

    So finallyyyy the MEGAAAA BAMMMMM is included.... Awesomeee

    • @statquest
      @statquest  5 ปีที่แล้ว +2

      Yes! I was hoping you would spot that! I did it just for you. :)

    • @rrrprogram8667
      @rrrprogram8667 5 ปีที่แล้ว +1

      @@statquest i was in office when i first wrote the comment earlier so couldn't see the full video...

  • @rungrawin1994
    @rungrawin1994 3 ปีที่แล้ว +2

    Listening to your song makes me thinking of Phoebe Buffay haha.
    Love it, anyway !

    • @statquest
      @statquest  3 ปีที่แล้ว +1

      See: th-cam.com/video/D0efHEJsfHo/w-d-xo.html

    • @rungrawin1994
      @rungrawin1994 3 ปีที่แล้ว +1

      ​@@statquest Smelly stat, smelly stat, It's not your fault (to be so hard to understand)

    • @rungrawin1994
      @rungrawin1994 3 ปีที่แล้ว

      @@statquest btw i like your explanation on gradient boost too

  • @123chith
    @123chith 5 ปีที่แล้ว +16

    Thank you so much can you please make a video for Support Vector Machines

  • @abdelhadi6022
    @abdelhadi6022 5 ปีที่แล้ว +1

    Thank you, awesome video

    • @statquest
      @statquest  5 ปีที่แล้ว

      Thank you! :)

  • @CC-um5mh
    @CC-um5mh 5 ปีที่แล้ว +1

    This is absolutely a great video. Will you cover why we can use residual/(p*(1-p)) as the log of odds in your next video? Very excited for the part 4!!

    • @statquest
      @statquest  5 ปีที่แล้ว +1

      Yes! The derivation is pretty long - lots of little steps, but I'll work it out entirely in the next video. I'm really excited about it as well. It should be out in a little over a week.

  • @joeroc4622
    @joeroc4622 4 ปีที่แล้ว +1

    Thank you very much for sharing! :)

  • @叶涵
    @叶涵 4 ปีที่แล้ว +1

    very detailed and convincing

    • @statquest
      @statquest  4 ปีที่แล้ว

      Thank you! :)

  • @sandipansarkar9211
    @sandipansarkar9211 2 ปีที่แล้ว +1

    finished watching

  • @SomeGuy-q1d
    @SomeGuy-q1d 4 ปีที่แล้ว +3

    Gradient Boost: BAM
    Gradient Boost: Double BAM
    Gradient Boost: Triple BAM
    Gradient Boost: Quadruple BAM
    Great Gradient Boost franchise)

    • @statquest
      @statquest  4 ปีที่แล้ว

      Thanks so much! XGBoost is next! It's an even bigger and more complicated algorithm, so it will be many, many BAMs! :)

    • @kunlunl6723
      @kunlunl6723 4 ปีที่แล้ว

      I thought you are ganna say PentaBAM -> Unstoppable -> Godlike (if you play League of Legend

  • @krazykidd93
    @krazykidd93 หลายเดือนก่อน +1

    Wow! I haven't seen a Mega BAM before!

    • @statquest
      @statquest  หลายเดือนก่อน

      :) That was for a special friend.

  • @tumul1474
    @tumul1474 5 ปีที่แล้ว +2

    amazing as always !!

    • @statquest
      @statquest  5 ปีที่แล้ว

      Any time! :)

  • @TechBoy1
    @TechBoy1 ปีที่แล้ว +1

    The legendary MEGA BAM!!

    • @statquest
      @statquest  ปีที่แล้ว

      Ha! Thank you! :)

  • @rodrigomaldonado5280
    @rodrigomaldonado5280 5 ปีที่แล้ว +4

    Hi Statquest would you please make a video about naive bayes? Please it would be really helpful

  • @shrilekhaa
    @shrilekhaa 3 หลายเดือนก่อน +1

    Awesome

    • @statquest
      @statquest  3 หลายเดือนก่อน

      Thanks!

  • @61_shivangbhardwaj46
    @61_shivangbhardwaj46 3 ปีที่แล้ว +1

    You r amazing sir! 😊 Great content

    • @statquest
      @statquest  3 ปีที่แล้ว

      Thanks a ton! :)

  • @rohitbansal3032
    @rohitbansal3032 3 ปีที่แล้ว +1

    You are awesome !!

  • @lakshman587
    @lakshman587 3 ปีที่แล้ว +4

    16:25 My first *Mega Bam!!!*

  • @yjj.7673
    @yjj.7673 5 ปีที่แล้ว +1

    This is great!!!

    • @statquest
      @statquest  5 ปีที่แล้ว

      Thank you! :)

  • @IQstrategy
    @IQstrategy 5 ปีที่แล้ว +1

    Great videos again! XGBoost next? As this is supposed to solve both variance (RF) & bias (Boost) problems.

  • @rvstats_ES
    @rvstats_ES 4 ปีที่แล้ว +1

    Congrats!! Nice video! Ultra bam!!

    • @statquest
      @statquest  4 ปีที่แล้ว

      Thank you very much! :)

  • @TheAbhiporwal
    @TheAbhiporwal 5 ปีที่แล้ว +2

    Superb video without a doubt!!!
    one query Josh, do you have any plans to cover a video on "LightGBM" in near future?

  • @pmanojkumar5260
    @pmanojkumar5260 5 ปีที่แล้ว +1

    Great ..

  • @sandralydiametsaha9261
    @sandralydiametsaha9261 5 ปีที่แล้ว +1

    thank you very much for your videos !
    when will you post the next one ?

  • @amanbagrecha
    @amanbagrecha 3 ปีที่แล้ว

    Need to learn how to run powerpoint presentation lol. Amazing stuff

  • @FelipeArayaable
    @FelipeArayaable 6 หลายเดือนก่อน +1

    Hello Josh! I think that there might be a mistake in methodology at min 5:11 compared to what you showed in part 4 of the series for computing the residual. In this video, the first set of residuals you computed it as (Observed - log(odds) = residuals) and in part 4 you calculate it as (Observed - probability = residuals), so in this scenario where we have Observed as 1, log(odds) as 0.7, and p as 0.66, shouldn't the residuals be (1 - 0.66 = 0.33) instead of (1 - 0.7 - 0.3)?
    Love your videos and I am a huge fan!

    • @statquest
      @statquest  6 หลายเดือนก่อน

      I think you are confused, perhaps because the log(odds) = log(4/2) = 0.7 = 4/6 = probability. So, in this specific situation, both the log(odds) and the probability are the same. Thus, when we calculate the residuals, we use the probability. The equation is Residual = (observed - probability), as can been see in earlier at 4:49

  • @mengdayu6203
    @mengdayu6203 5 ปีที่แล้ว +17

    How does the multi-classification algorithm work in this case? Using one vs rest method?

    • @questforprogramming
      @questforprogramming 4 ปีที่แล้ว +2

      It's been over 11 months and no reply from josh... bummer

    • @AnushaCM
      @AnushaCM 4 ปีที่แล้ว +2

      have the same question

    • @ketanshetye5029
      @ketanshetye5029 4 ปีที่แล้ว +1

      @@AnushaCM well, we could use one vs rest approach

    • @Andynath100
      @Andynath100 4 ปีที่แล้ว +3

      It uses a Softmax objective in the case of multi-class classification. Much like Logistic(Softmax) regression.

  • @stephentan9523
    @stephentan9523 2 ปีที่แล้ว +1

    That's a rare mega bam!!! I can die peacefully now.

  • @ErnieHsieh-f4j
    @ErnieHsieh-f4j 7 หลายเดือนก่อน

    Best video ever, quick question on building the next tree. Once we have the new residuals, how do we decide the new node for the next tree? Is it still the same as calculating Gini but on the residuals ?

    • @statquest
      @statquest  7 หลายเดือนก่อน

      Gradient Boost traditionally uses Regression Trees. If you don't already know about Regression Trees, check out the 'Quest: th-cam.com/video/g9c66TUylZ4/w-d-xo.html

  • @deepakmehta1813
    @deepakmehta1813 3 ปีที่แล้ว

    Fantastic song, Josh. I have started picturing that I am attending a class and the professor/lecturer walks by in the room with the guitar, and the greeting would be the song. This could be the new norm following stat quest. One question regarding gradient boost that I have is why it restricts the size of the tree based on the number of leaves. What would happen if that restriction is ignored? Thanks, Josh. Once again, superb video on this topic.

    • @statquest
      @statquest  3 ปีที่แล้ว

      If you build full sized trees then you would overfit the data and you would not be using "weak learners".

  • @questforprogramming
    @questforprogramming 2 หลายเดือนก่อน

    1. Calculate logodds.
    2. Convert logodds to calculate probabilities.
    3. Calculate residuals.
    4. Build tree.
    5. Calculate output values per leaf in tree.
    6. Calculate 2nd log odds using first logodds + learning rate * predictions from tree.
    7. Convert 2nd logodds to calculate probabilities.
    8. Calculate 2nd residuals.
    REPEAT...

  • @KalyanGk0
    @KalyanGk0 3 ปีที่แล้ว +1

    This guy literally coming to my dreams 😂

  • @ulrichwake1656
    @ulrichwake1656 5 ปีที่แล้ว +3

    Thank you so much. Great videos again and again.
    One question, what is the difference between xgboost and gradient boost?

    • @mrsamvs
      @mrsamvs 4 ปีที่แล้ว

      please reply @statQuest team

  • @enkhbaatarpurevbat3116
    @enkhbaatarpurevbat3116 4 ปีที่แล้ว +1

    love it

  • @jayyang7716
    @jayyang7716 2 ปีที่แล้ว +1

    Thanks so much for the amazing videos as always! One question: why the loss function for Gradient Boost classification uses residual instead of cross entropy? Thanks!

    • @statquest
      @statquest  2 ปีที่แล้ว

      Because we only have two different classifications. If we had more, we could use soft max to convert the predictions to probabilities and then use cross entropy for the loss.

    • @jayyang7716
      @jayyang7716 2 ปีที่แล้ว

      @@statquest Thank you!

  • @JoaoVictor-sw9go
    @JoaoVictor-sw9go 2 ปีที่แล้ว +2

    Hi Josh, great video as always! Can you explain to me or recommend a material to understand the GB algorithm when we are using it for a non-binary classification? E.g. we have three or more possible outputs for classification.

    • @statquest
      @statquest  2 ปีที่แล้ว +1

      Unfortunately I don't know a lot about that topic. :(