Introduction to Random Forest | Intuition behind the Algorithm

แชร์
ฝัง
  • เผยแพร่เมื่อ 24 ม.ค. 2025

ความคิดเห็น • 69

  • @priyankamahamuni-d3l
    @priyankamahamuni-d3l ปีที่แล้ว +9

    you are great teacher....great data scientist....hats off to you....because of you all the basic , small n small concepts getting cleared...you are faaabbb

  • @rajatchauhan4410
    @rajatchauhan4410 ปีที่แล้ว +10

    Hi, thanks for the great explanation but just a silly mistake I think after we did column sampling, we will need to give data point for prediction in same order as the features selected at 30:00

  • @devamsingh2352
    @devamsingh2352 4 หลายเดือนก่อน +12

    completed 90 videos with notes within 3 weeks 🥳🥳🥳🥳. I still remember all the topics and working. Your teachings are blessings from god.

    • @Sahil-xy2tp
      @Sahil-xy2tp 2 หลายเดือนก่อน

      Bro could you please share that notes

    • @sourabh_321
      @sourabh_321 2 หลายเดือนก่อน

      Yes can u pls provide notes

    • @hardik3809
      @hardik3809 2 หลายเดือนก่อน

      @@sourabh_321 I have also prepared the notes. But those includes only theory and maths portion, not the coding part.

    • @sourabh_321
      @sourabh_321 2 หลายเดือนก่อน +2

      @@hardik3809 I don't need coding part I need theoretical part if u can provide me this will help me a lot ❤️

    • @PRATHAMBHUSHANGHILORIA
      @PRATHAMBHUSHANGHILORIA 2 หลายเดือนก่อน +1

      bhai bhejde paper aa rhe h

  • @siyays1868
    @siyays1868 2 ปีที่แล้ว +10

    I m a great fan of u & ur channel. Thanku so much for working hard. Itna mast Random forest shayad hi kisine explain kiya hoga youtube par. Not only Random Forest actually all other algorithms or different concepts in Data science , all videos of urs r the best of the best.

  • @subhendukundu4595
    @subhendukundu4595 6 หลายเดือนก่อน +5

    You deserve 1M Subscribers!!

  • @dikshakushwaha8282
    @dikshakushwaha8282 หลายเดือนก่อน

    you Truely deserve more than 1M followers. mei apki bahut badi fan hu sir hu dil se, app jb padhate ho to padhayi se pyaar sa ho jata h.

  • @siyays1868
    @siyays1868 2 ปีที่แล้ว +2

    Best ever channel & best ever teacher for data science . Thanku very very much.

  • @bhupendersharma0428
    @bhupendersharma0428 2 ปีที่แล้ว +3

    It is best Channal for data Science It will grow 1M in 2023 i swear

    • @lakshaychauhan380
      @lakshaychauhan380 ปีที่แล้ว +1

      Nhi hue bhai🙃

    • @shashibhushanjha7325
      @shashibhushanjha7325 9 หลายเดือนก่อน +3

      Its good that less but genuine people are following this channel. Knowledge is for everyone but only dedicated minds will make use of it, other just follow their adrenaline rush.

  • @aounhaider8335
    @aounhaider8335 ปีที่แล้ว +4

    Your videos on ML are amazing. Following this playlist!!

    • @Human12358
      @Human12358 ปีที่แล้ว +1

      But little bit long

    • @sanjaisrao484
      @sanjaisrao484 ปีที่แล้ว

      ​@@Human12358yes

    • @indra-zd9zu
      @indra-zd9zu 5 หลายเดือนก่อน

      But worthy​@@Human12358

  • @_pareekshithmcMcpareekshith
    @_pareekshithmcMcpareekshith 10 หลายเดือนก่อน

    the way you teach is absolutely amazing , keep up the good work ,thankyou

  • @anilkathayat1247
    @anilkathayat1247 9 หลายเดือนก่อน

    Your explanation on each small point is next level! Great job sir.

  • @sourabhyadav8258
    @sourabhyadav8258 9 หลายเดือนก่อน

    Great content totally diffrent way of teaching!! Mark my word nobody spent so much time on a Single project but here the story is diffrent!!

  • @arpittrivedi6636
    @arpittrivedi6636 2 ปีที่แล้ว +1

    Bahut badiya sir. God bless you 🙏🙏

  • @rahmankhan7303
    @rahmankhan7303 หลายเดือนก่อน

    you are thanos for complex topics

  • @studology67
    @studology67 28 วันที่ผ่านมา

    In column sampling prediction, How can we give same data to each decision trees as there can be the case that all decision tree have different columns

  • @johnson2784
    @johnson2784 ปีที่แล้ว

    massive respect ❤,
    you are a great teacher

  • @user-bt6mh9ez3u
    @user-bt6mh9ez3u 6 หลายเดือนก่อน

    if we do not specify any sampling ,by default is it taking hybrid(row and column sampling together)like some rows and feature(column) are taken together for training decision tree?

  • @vishnuvardhanjadava4186
    @vishnuvardhanjadava4186 11 หลายเดือนก่อน +2

    there is a small mistake in the column_row sampling(last one). df1,2,3 have different features which are trained on different models and while performing the prediction, you passed he same features for all three models and took majority(aggregation). apart from this rest is awesome.
    I have a question. lets say I run the random forest by giving my dataset with columns from F1 to F10 with target variable Y. I ran some sklearn model or something. i did train test split and i did fit and now, I want to perform prediction. since, each DT model in RF has different and only few cols(let's say 50%) and my test data will have all the cols. Are the test data features gonna be passed to respective DT model with respective features that the model was trained on? or is there some other mechanism? please explain.

    • @saikrishna-p9c
      @saikrishna-p9c 6 หลายเดือนก่อน

      same doubt bro .comment if u get the answer

    • @caul69
      @caul69 4 หลายเดือนก่อน

      All the features are gonna be passed to the DT models, the classification is done by each model using the subset of features they were trained on , which is then aggregated and you get the RandomForest prediction

  • @SourabhGupta108
    @SourabhGupta108 ปีที่แล้ว +7

    I think there is some mistake while using column sampling, in a prediction part you are passing the same input array to all the decision trees while it should be different according to the sampled dataframe.

    • @Shubham_gupta18
      @Shubham_gupta18 8 หลายเดือนก่อน

      yes, it would be

    • @amolhire9482
      @amolhire9482 7 หลายเดือนก่อน

      yes we cant predict values like that if we have input data scaled with diff. features , yes sir se thoda mistake hogaya hain.

  • @minalgupta7456
    @minalgupta7456 8 หลายเดือนก่อน

    I m a great fan of u & ur channel

  • @narendraparmar1631
    @narendraparmar1631 ปีที่แล้ว

    Very Well Explained
    Thanks Sir

  • @balrajprajesh6473
    @balrajprajesh6473 2 ปีที่แล้ว

    best of the best!

  • @sarveshjoshi2611
    @sarveshjoshi2611 15 วันที่ผ่านมา

    I am glad he is underrated.

    • @deeksha-cm8kq
      @deeksha-cm8kq 4 วันที่ผ่านมา

      why are you glad then? he should and will get more popular.

  • @rockykumarverma980
    @rockykumarverma980 3 หลายเดือนก่อน

    Thank you so much sir 🙏🙏🙏

  • @amolhire9482
    @amolhire9482 7 หลายเดือนก่อน

    sir hum random features selection karenge aur us par value prediction bhi karenge dt se lekin agar hume testing / input data whole features ka mila toh prediction kaise hoga ? i think row sampling is fine but features sampling not understood .

  • @brayanrai2880
    @brayanrai2880 ปีที่แล้ว

    Best one

  • @taslima5007
    @taslima5007 10 หลายเดือนก่อน

    You are great

  • @souravaich6620
    @souravaich6620 2 ปีที่แล้ว

    I am still confused on sampling without replacement, if we have 10 features and we selecting let say 5 features for each DT, then my 10 features will be exhausted in 2 DTs only. Then how come it's distributing in i.e. 100 DTs? Same with row sampling we are giving 25% of rows to each DTs then it should be exhausted after 4 DTs, then how come we are training the other DTs. Please help me with this, I am totally confused with 'without replacement' option.

    • @aazeebh3734
      @aazeebh3734 2 ปีที่แล้ว +2

      In sampling without replacement:
      For the 1st tree, if 5 columns are to be selected: For the 1st column, all 10 columns are available for selection. After the 1st column is selected, then the next column could be any of the remaining 9 columns.. So, this avoids repeating of the columns.
      For the next DT, the process starts from scrath. All the 10 columns are available for selection and the same process as above follows.
      In sampling with replacement, after the 1st column is selected, while selecting the 2nd column, it will have all the 10 columns to choose from, including the column already selected as 1st column.

  • @patelpavan5479
    @patelpavan5479 ปีที่แล้ว

    ca we take two output in RF

  • @sandipansarkar9211
    @sandipansarkar9211 2 ปีที่แล้ว

    finished watching

  • @vishalkumar-us7ys
    @vishalkumar-us7ys 3 ปีที่แล้ว

    While teaching preprocessing you mentioned that we check for duplicate data (rows) and drop it as it can create problem while building model then how random forest handles the duplicate data which comes from sampling with replacement techniques.

    • @campusx-official
      @campusx-official  3 ปีที่แล้ว +5

      Good question, Atul. See, the problem with duplicate data is that it increases model bias which may lead to overfitting. But the way RF works, it is able to handle this bias.

  • @fgg102
    @fgg102 3 ปีที่แล้ว

    Hi, i have a small question about decision trees. Is it ok to have a decision tree with a max depth of 7, bcs i noticed how it produced a big tree but also had the better accuracy training and test score than if I had reduced the max depth.

    • @campusx-official
      @campusx-official  3 ปีที่แล้ว +1

      Yes you can have a decision tree of max depth 7 or more. It depends on the data

    • @fgg102
      @fgg102 3 ปีที่แล้ว +1

      @@campusx-official oh okay, both my train and test data show to be ok in terms of not over or under fitting !
      Btw thank you sir, your video literally saved my life! Bless you

  • @ABHISHEKKUMAR-gv4di
    @ABHISHEKKUMAR-gv4di 2 ปีที่แล้ว

    I am not able to find the link for the visualisation tool

  • @ashutoshthokare2127
    @ashutoshthokare2127 11 หลายเดือนก่อน

    Thank u sir

  • @RohitKumar-wb4pe
    @RohitKumar-wb4pe 7 หลายเดือนก่อน

    column sampling function not working .
    can anybody resolve it ?

  • @luffy6761.
    @luffy6761. ปีที่แล้ว +1

    sir please provide notes for this ml playlist

  • @AbdulRahman-zp5bp
    @AbdulRahman-zp5bp 3 ปีที่แล้ว

    THANK YOU 3000

  • @kindaeasy9797
    @kindaeasy9797 11 หลายเดือนก่อน

    Wow

  • @tanb13
    @tanb13 2 ปีที่แล้ว +1

    I think the explanation of bootstrapping is wrong in the video. As per Wikipedia (en.wikipedia.org/wiki/Bootstrap_aggregating) , in bootstrapping process the size of the sample dataset has to be same as the original dataset and it is not a smaller subset of the original dataset as explained in the video. I think what you have explained (creating smaller samples from the original dataset) is concept of bootstrapping but applicable from statistics standpoint and not in context of Random Forests. Still I will be grateful if you can kindly explain your reasoning with some related articles.

  • @purushottammitra1258
    @purushottammitra1258 3 ปีที่แล้ว

    How come 65th video after 62nd(ensemble technique) ??

  • @SatyamBonaparte
    @SatyamBonaparte 10 หลายเดือนก่อน +1

    This video was a bit confusing ngl

  • @hamzalaalj
    @hamzalaalj 11 หลายเดือนก่อน

    Since you're speaking Indian, at least write the title in Indian!!! So we avoid wasting our time!

    • @indra-zd9zu
      @indra-zd9zu 5 หลายเดือนก่อน

      It's his choice.bkl