Ridge Regression Part 1 | Geometric Intuition and Code | Regularized Linear Models

แชร์
ฝัง
  • เผยแพร่เมื่อ 26 ม.ค. 2025

ความคิดเห็น • 60

  • @mdnoorsolaimansiam8821
    @mdnoorsolaimansiam8821 2 ปีที่แล้ว +24

    For past two days, I was watching different videos and reading articles to understand the core of ridge regression. I got tired as I wasn't understanding. And here I am, after half past this video, I think I've got the grasp. It'd be biased to say previous contents didn't help me at all, but your lecture is so much insightful than those. Thank you very much for sharing your learnings with us.

  • @TomatoPotato874
    @TomatoPotato874 9 หลายเดือนก่อน +3

    Charansparsh aapko 🙏

  • @UtkarshSingh-qw6md
    @UtkarshSingh-qw6md 2 หลายเดือนก่อน

    Never ever have I seen such brilliance in teaching.

  • @quicksciencecalculation5676
    @quicksciencecalculation5676 2 ปีที่แล้ว +6

    Well explained sir please i request you never Stop

  • @youbeyou9769
    @youbeyou9769 ปีที่แล้ว +1

    Omg this is brilliant. Exactly what I've been looking for. Thanks for making our lives easier

  • @1234manasm
    @1234manasm 2 ปีที่แล้ว +1

    You are look like a genius and tech like a professor

  • @yuktashinde3636
    @yuktashinde3636 2 ปีที่แล้ว +1

    at 10:51 why we considered training datapoints in testing but yes there are other training points who will create y-y^ value

  • @23injeela79
    @23injeela79 8 หลายเดือนก่อน

    very well explained . sir you have done a lot of hardwork on your lectures . keep going .

  • @mustafizurrahman5699
    @mustafizurrahman5699 ปีที่แล้ว

    Mesmerising such lucid explanation

  • @arpitakar3384
    @arpitakar3384 5 หลายเดือนก่อน

    God of ML teaching

  • @ParthivShah
    @ParthivShah 10 หลายเดือนก่อน +1

    Thank You Sir.

  • @krishnakanthmacherla4431
    @krishnakanthmacherla4431 2 ปีที่แล้ว +1

    You are a game changer sir

  • @sujithsaikalakonda4863
    @sujithsaikalakonda4863 2 ปีที่แล้ว +2

    Great explanation sir.

  • @Anjalisharma-wp2tg
    @Anjalisharma-wp2tg หลายเดือนก่อน

    Nitish!!!! You are amazing

  • @RAKESHADHIKARI-h4b
    @RAKESHADHIKARI-h4b 6 หลายเดือนก่อน

    Awesome series brother. Great work done by you..Looking for Mlops video

  • @manasyadav3159
    @manasyadav3159 ปีที่แล้ว +2

    Hi, how can we say that if the slope is very high then it will be the case of overfitting, it can be underfitting also. I think high slope doesn't mean it will perfectly fit on our training data. Please help me out.

  • @rahmankhan7303
    @rahmankhan7303 2 หลายเดือนก่อน

    this is the first video whose concept was'nt very cleared for me but still great video sir

  • @ghostofuchiha8124
    @ghostofuchiha8124 11 หลายเดือนก่อน +1

    Krish Naik hindi has explained this better ; Rest others till now ; CampusX seems good

    • @darshedits1732
      @darshedits1732 11 หลายเดือนก่อน

      For only this algorithm krish naik is explained better or for all algorithms krish naik are explained better?

    • @SubhamKumar-bt9th
      @SubhamKumar-bt9th 4 หลายเดือนก่อน

      i watched his video just now after reading your comment, it's mostly same nothing better, even he has not explained why, "what if the incorrect fit line is already on the right side and imaginary true fit is on left ,then ridge will shift it more right away from true fit,, ".?"

  • @UtkarshSinghchutiyaNo1
    @UtkarshSinghchutiyaNo1 2 หลายเดือนก่อน

    Sir Legend ho aap

  • @krishcp7718
    @krishcp7718 ปีที่แล้ว +1

    Hi Nitish,
    Very nice video.
    Just one thing I noticed - around 04:00. For a given intercept b, when m changes, it is basically the orientation or the angle the line makes with the x-axis changing. So when m is either too small or too high, there is underfitting. As can be seen geometrically the line is quite away from the data points for high and low m. So the overfitting - meaning line is very close to the data points is for only certain values of m - particularly between high m and low m values. Please let me know your thoughts on this.
    Regards,
    Krish

    • @Ishant875
      @Ishant875 ปีที่แล้ว

      That statement is incorrect

  • @ATHARVA89
    @ATHARVA89 3 ปีที่แล้ว +5

    Sir will you including SVM t-sne and all ahead in the 100 days ML playlist?

  • @shashankbangera7753
    @shashankbangera7753 ปีที่แล้ว

    Beautiful explaination!

  • @osho_magic
    @osho_magic 2 ปีที่แล้ว +1

    Sir what if the incorrect fit line is already on the right side and imaginary true fit is on left ,then ridge will shift it more right away from true fit,, ".? It become irregularisation. Isn’t it.?

  • @harsh2014
    @harsh2014 2 ปีที่แล้ว

    Thanks for this session.

  • @Ishant875
    @Ishant875 ปีที่แล้ว +1

    I appreciate your work and no one can teach you like you but there is just a thing, Overfitting doesn't mean high slope in simple linear regression. Overfitting means you have used very complex model which is not able to generalise for new data which is not in training. Simple linear regression is simplest model, so they can't be overfitting in it. There can be only underfitting.

    • @kindaeasy9797
      @kindaeasy9797 8 หลายเดือนก่อน

      exactly , simple linear regression mai over fitting ho hi nahi skti , because the line will not bend to pass from each and every data point in the training data set , yes it can underfit and best fit

  • @arman_shekh97
    @arman_shekh97 3 ปีที่แล้ว

    Sir this video has cames after 5 days , everything is fine now

  • @BP-me7lj
    @BP-me7lj 4 หลายเดือนก่อน

    The whole idea should be to reduce the overfitting of 1st line. But we are having 2nd linewith different parameters. There should be only first line and when we multiply it with lambda*m^2 then it should give less error. Here we already having 2nd line. When i calculated the loss without lambda term loss was even less. Idk. Someone please clearify this.

  • @kindaeasy9797
    @kindaeasy9797 8 หลายเดือนก่อน

    but in the graph at 7:28 if jo 2 training points hai vo test points ke neeche hotai and best fit line ke slope ko increase krna padta na overfitting ko handle krne ke liye ??? I think is video ka logic flawed hai

  • @d-pain4844
    @d-pain4844 2 ปีที่แล้ว +1

    Sir thoda dark marker use Karo

  • @asifkdas
    @asifkdas ปีที่แล้ว

    the slope value in a linear regression model does not directly indicate overfitting

    • @Tusharchitrakar
      @Tusharchitrakar ปีที่แล้ว

      Yes ofcourse but I think what he's trying to suggest is that some suspiciously high values "might" be indicative of over fitting.

  • @lingasodanapalli615
    @lingasodanapalli615 9 หลายเดือนก่อน

    But sir why did you choose two training points above the actual dataset. If we chose those two training points below the actual dataset then the correct line's slope is higher than predicted lin's slope. So the loss of the predicted line's slow will be less

    • @ruchitkmeme4441
      @ruchitkmeme4441 8 หลายเดือนก่อน

      exactly for this problem i came into comment box!!
      i mean if we give all data and not only that two point normal linear regretion will also choose that line that we want after ridge regression

  • @shaikhsaniya3585
    @shaikhsaniya3585 2 ปีที่แล้ว

    Regularization is regression here ??

  • @abhinavkale4632
    @abhinavkale4632 3 ปีที่แล้ว +1

    Just one issue.. why did you multiply 0.9*3 while calculating the loss at second point?

    • @SidIndian082
      @SidIndian082 2 ปีที่แล้ว +1

      even i am confused on this .,..:-(

    • @balrajprajesh6473
      @balrajprajesh6473 2 ปีที่แล้ว +1

      It is clearly mentioned by Sir in the video that it is just an assumption.

    • @somanshkumar1325
      @somanshkumar1325 2 ปีที่แล้ว +2

      There are two points in our training dataset -> (1,2.3) and (3,5.3).
      For calculating the loss at the second point,
      Yi = 5.3, Xi = 3.
      Y_hat = m*Xi + b where m=0.9, Xi = 3, b=1.5.
      Y_hat = 0.9*3+1.5
      I hope it helps?

    • @abhinavkale4632
      @abhinavkale4632 2 ปีที่แล้ว

      @@somanshkumar1325 yooo...

  • @mohitkushwaha8974
    @mohitkushwaha8974 2 ปีที่แล้ว

    Awesome

  • @ANUbhav918
    @ANUbhav918 3 ปีที่แล้ว

    I guess, Bias is more compared to variance in Overfitting. Vice versa in Underfitting. Please correct me

  • @kindaeasy9797
    @kindaeasy9797 8 หลายเดือนก่อน

    aapki jo overfitting wali line hai , vo toh low bias and high variance wali definition ko satisfy hi nahi kr rhi , i dont think that overfitting is possible in case of simple linear regg, because the line cant bend to pass from each and every data point of the training data set

  • @pankajwalia155
    @pankajwalia155 2 ปีที่แล้ว

    Koi please bata do training error generalized error testing error irreducible error kis section main hai mera exam hain 20 dec ko

  • @rohitdahiya6697
    @rohitdahiya6697 2 ปีที่แล้ว

    why there is no learning rate hyperparameter in scikit-learn Ridge/lasso/Elasticnet . As it has a hyperparameter called max_iteration that means it uses gradient descent but still there is no learning rate present in hyperparameters . if anyone knows please help me out with it.

    • @near_.
      @near_. 2 ปีที่แล้ว

      Did u get the answer??

    • @rohitdahiya6697
      @rohitdahiya6697 2 ปีที่แล้ว

      @@near_. no still waiting for some expert to reply

    • @YogaNarasimhaEpuri
      @YogaNarasimhaEpuri 2 ปีที่แล้ว +1

      I didn't thought about this...
      Just I seen from documentation, that All Solvers are not using Gradient Descent. (which i think)
      SAG - uses a Stochastic Average Gradient descent, The (step size/learning rate) is set to 1 / (alpha_scaled + L + fit_intercept) where L is the max sum of squares for over all samples.
      ‘svd’ uses a Singular Value Decomposition, (Matrices)
      cholesky, (Matrices)
      ...
      Otherwise, Like SAG, all solvers based upon the data and solver, automatically calculate the learning rate
      What's ur opinion?

  • @arshad1781
    @arshad1781 3 ปีที่แล้ว

    THANK

  • @garimadhanania1853
    @garimadhanania1853 8 หลายเดือนก่อน

    I really appreciate your effort and all your videos but I think the explanation is incorrect here.
    m being high is not the definition of overfitting
    in a typical linear regression with m + 1 weights, if we do not constrain the value of weights and let them be anything, then they can represent very complex functions and that causes overfitting
    we have to penalize large values of weights (by adding in the loss function) so that our function has lower capacity to represent complexity and hence it wont learn complex functions that just fit the training data well

    • @aaloo_ka_paratha
      @aaloo_ka_paratha 6 วันที่ผ่านมา

      You're absolutely right. Overfitting occurs when the model becomes too complex, which can happen if the weights are unconstrained and grow too large, allowing the model to fit the noise in the data. Regularization techniques like Ridge regression help prevent this by adding a penalty to the weights, ensuring the model remains simpler and generalizes better to unseen data. Great explanation!

  • @ashisharora9649
    @ashisharora9649 2 ปีที่แล้ว

    Sorry to say but you share the bookish knowledge this time. Practical intuition is not there. Adding something parallelly shifts the line upward. How does it able to make change in the slope? You said, mostly ds keeps this model as default as it will only be active if there is a situation of overfitting, Kindly explain that how? How model is able to find the best fit line of test set is that you assumed it on your own. Does algorithm do the same?

    • @campusx-official
      @campusx-official  2 ปีที่แล้ว

      Regularization: th-cam.com/play/PLKnIA16_RmvZuSEZ24Wlm13QpsfLlJBM4.html
      Check out this playlist, maybe this will help

  • @rk_dixit
    @rk_dixit 4 หลายเดือนก่อน

    plz spend a lot more time on code