3.5: Mathematics of Gradient Descent - Intelligence and Learning

แชร์
ฝัง
  • เผยแพร่เมื่อ 26 ม.ค. 2025

ความคิดเห็น • 485

  • @vcrmartinez
    @vcrmartinez 6 ปีที่แล้ว +38

    Hey, I'm a Brazilian student/software engineer studying Rec Systems and ML. Tons of articles, papers and videos did not do what you've just done. Now everything is crystal clear, thanks for the explanation.

  • @stevenrhodes2818
    @stevenrhodes2818 6 ปีที่แล้ว +91

    OMG! Thank you! My power went out and I figured I would try to learn gradient descent on my phone.. This is the first time it's made sense.. All those experienced mathmaticians suck at being teachers, making it sound all crazy complicated and shit. You sir are amazing.

    • @fernandoerazo8135
      @fernandoerazo8135 6 หลายเดือนก่อน

      My wifi just went out and instead of using it as an excuse, I am using my laggy phone to try and learn .

  • @JulietNovember9
    @JulietNovember9 6 ปีที่แล้ว +139

    Going through Andrew Ng's Coursera... got stuck on how the Cost Function derivatives/partial derivatives are obtained.... 11:00 and on... Oh... MY... GOSH... this is GOLD!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! Thank you so much!

    • @calluma8472
      @calluma8472 5 ปีที่แล้ว +3

      Same. Haven't done calculus since 1999, this helps a lot.

    • @seamusforever7081
      @seamusforever7081 5 ปีที่แล้ว

      Same here. I was so confused.

    • @tharukabalasooriya3269
      @tharukabalasooriya3269 5 ปีที่แล้ว

      @@calluma8472 gosh, how old are you?

    • @abirchakraborty624
      @abirchakraborty624 5 ปีที่แล้ว +1

      Spot on ... This is a small bridge for the Andrew Ng's Coursera course . Specially the explanation how chain rule and power rule are coming into picture here, really helps.

    • @pemadechen9901
      @pemadechen9901 4 ปีที่แล้ว +1

      found my coursera classmate

  • @iamgrinhausgases
    @iamgrinhausgases 8 หลายเดือนก่อน +1

    Just wanted to say that this is easily the best and clearest explanation of gradient descent I've come across, on the web and in the books I've read. Thank you, sir.

  • @nikitachoudhary3878
    @nikitachoudhary3878 5 ปีที่แล้ว +1

    after read 20-30 articles , after watched 20+ videos, i watched best video ever of Gradient Descent.
    i falling love with you.
    best explanation.

  • @SaurabhSinghAus
    @SaurabhSinghAus 5 ปีที่แล้ว +5

    Coming here from Andrew Ng's ML course. Got confused with Gradient Descent. This is Gold. You explained Linear regression so well.

  • @alishaaluu
    @alishaaluu 5 ปีที่แล้ว +25

    Usually, math makes me cry but while watching this I am learning and laughing at the same time. How cool is that? Lol. All thanks to you, bro! Keep the good work on. Cheers!!

  • @ajitkumar15
    @ajitkumar15 6 ปีที่แล้ว +1

    Thank you so much.After failing my exam on Machine learning I was searching videos on Gradient Descent topic.After watching so many videos I landed on this page.By far this is the best video.You are simple great teacher because you have understood the topic very well that is why you are able to explain in really simple way.. Thanks a million !!!!!

  • @TheSocialDrone
    @TheSocialDrone 4 ปีที่แล้ว +2

    I had searched on this subject and watched several other videos before I could find this amazing video on the topic; I am more than happy that I am now able to explain this concept to anyone - now it is so much clear, thank you sir!

  • @umessi10
    @umessi10 7 ปีที่แล้ว +3

    Thank you so much for this. One of the best explanation of gradient descent on youtube. So far I'm loving your Intelligence and Learning series. Think I'm gonna binge watch the entire series now.

  • @TimHulse
    @TimHulse 4 ปีที่แล้ว +1

    You nailed it sir! I was confused when partial dJ/dm = 2*Error at one moment then suddenly had partial dError/dm glued onto it, but your clarification at point 19:15 clarified it. Please keep making videos!

  • @badcode3976
    @badcode3976 4 ปีที่แล้ว

    I have seen some of your videos to get some concepts that I didn't get the first time on my ML class and I'm truly convinced that these are the best tutorials on TH-cam about ML, you make every concept so simple to understand and funny, at the same time. Thanks a lot!!! Keep doing this great content!!!

  • @CamdenBloke
    @CamdenBloke 6 ปีที่แล้ว +7

    I'm studying up for an interview to transfer to the Machine Learning department Wednesday. This is enormously helpful in providing an actual mathematical (not just conceptual) understanding of gradient descent. Thanks!

  • @shivamkaushik6637
    @shivamkaushik6637 6 ปีที่แล้ว +1

    It was teacher's day yesterday, here in India. And today I have got this amazing teacher. Thank You

  • @r.d.machinery3749
    @r.d.machinery3749 5 ปีที่แล้ว

    This is a clearer explanation than Professor Ng's explanation in his machine learning video series. Ng denotes m and b as theta0 and theta1. He also reverses the terms in his line equation which confuses the Hell out of everybody. In addition, he doesn't take you through how the partial derivative is worked out and he doesn't show the code. A great explanation in only 22 minutes.

  • @watchnarutoshippuden3228
    @watchnarutoshippuden3228 6 ปีที่แล้ว

    Your videos are the only calculus and ML videos I can understand, you are the best! I just subbed to you; 'minimum' is singular and 'minima' is plural.

  • @matheuscorrea3277
    @matheuscorrea3277 4 ปีที่แล้ว

    I searched for a video like this for a long time, and the only one I could clearly understood was yours. thank you so much and congrats for the explanation

  • @simaosoares165
    @simaosoares165 7 ปีที่แล้ว

    You did it fantasticly! These are concepts that I know well already but find them difficult to explain, so I'll recommend your videos when pure IT guys (and not so educated audiences) ask me about the internals of the ML algos that I use.

  • @paedrufernando2351
    @paedrufernando2351 6 ปีที่แล้ว

    you deserve 200 million subscribers.. more than that your personality is really great!!!

  • @rupambanerjee3
    @rupambanerjee3 5 ปีที่แล้ว +1

    This is the best gradient descent video I have ever seen! Great work!

  • @lggood8375
    @lggood8375 4 ปีที่แล้ว

    Thanks for doing this within the 20 minutes you made it clear. Than the many hours I have watched and read articles others have made and were totally confusing. keep doing this... you made me unafraid of all the math.

  • @BinuVADAT
    @BinuVADAT 6 ปีที่แล้ว

    You went in to details and explained the concept. Loved the way of your fun-filled teaching. Thank you !

  • @8eck
    @8eck 4 ปีที่แล้ว

    3Blue1Brown's was hard for me, your explanations are waaay better.

  • @alexmisto423
    @alexmisto423 5 ปีที่แล้ว

    After watching this i finally figured out the calculus behind back propagation. Thank you! BIG LIKE

  • @-long-
    @-long- 6 ปีที่แล้ว

    I never saw any tutor who put so much emotion in the video like you lol
    Excellent channel! Thanks so much

  • @via_domus
    @via_domus 6 ปีที่แล้ว +43

    you're a very good teacher, a bit crazy though lol

    • @Dennis4Videos
      @Dennis4Videos 4 ปีที่แล้ว +6

      The great kind of crazy tho!

    • @anusha8085
      @anusha8085 3 ปีที่แล้ว +1

      @@Dennis4Videos yes!

  • @cuteworld9310
    @cuteworld9310 3 ปีที่แล้ว

    Amazing esplanation, easy enough for a high school student to learn. Amazing how simple you made this complex concept. You sir are a genius!

  • @timt.4040
    @timt.4040 7 ปีที่แล้ว

    Was looking for an accessible explanation of gradient descent, and this was by far the best one I found--thanks!

  • @smrazaabidi1495
    @smrazaabidi1495 7 ปีที่แล้ว +2

    Really awesome, what a elegant style of delivering the concepts, mind boggling. I wish & dream i should work and get education under his supervision. Moreover, gestures, tone, humor was extra extra outstanding, i'm speechless. :). I must say that it is the best ever explanation of gradient descent I've seen so far. Thanks a lot.

  •  7 ปีที่แล้ว

    I can not believe this video aired just when i needed it, that you so much!

  • @abirchakraborty624
    @abirchakraborty624 5 ปีที่แล้ว

    Spot on ... This is a small (but a very important ) bridge for the Andrew Ng's Coursera course . Specially the explanation how chain rule and power rule are coming into picture here, really helps.

  • @feliciafryer3271
    @feliciafryer3271 7 ปีที่แล้ว

    Thank you!!! I’m a computer science graduate student and trying to understand gradient descent. This video is awesome, can’t wait to watch more of your videos.

  • @LudwigvanBeethoven2
    @LudwigvanBeethoven2 6 ปีที่แล้ว

    You are so engaging that turns this boring math to something actually interesting.. thank you so much

  • @JotaFaD
    @JotaFaD 7 ปีที่แล้ว

    A great youtuber recommending another one! Although I know Calculus from college, I think you did a great job explaining some of the rules. Keep it up Daniel.

  • @azr_sd
    @azr_sd 5 ปีที่แล้ว +3

    Daniel I have been following you since you had 2000 subs. I always enjoyed your videos man. I started learning Deep Learning on my own and got stuck at understanding Gradient descent and I know it is the back bone of Ml and DL I want to know it deeply. I have watched around 3 videos before this and your video just explains it beautifully. Thanks for this video it helped me alot. Please keep doing these kind of videos which explains the math behind these ML and DL algorithms and again Thank you for your videos. :) Iam gonna follow you more and more from now. If it's possible try to make an awesome course on Udemy with math and programming of ML and DL . Thank you again.

  • @swatigautam9802
    @swatigautam9802 7 ปีที่แล้ว

    You're videos are entertaining and informative at the same time . Love it!

  • @praveenharris6170
    @praveenharris6170 6 ปีที่แล้ว

    I cannot describe how useful this was to me! Thank you!

  • @TheBestcommentor
    @TheBestcommentor 5 ปีที่แล้ว +1

    I've been studying this subject for a couple months in my final semester of college, and for reason, the connectiom between the loss function and the parabola just made it all click

  • @spacedustpi
    @spacedustpi 6 ปีที่แล้ว +1

    This was the best tutorial on this subject that I've found, thank you for this too!

  • @ryanmccauley211
    @ryanmccauley211 7 ปีที่แล้ว

    After watching a ton of videos I finally understand it thanks to you. Thank you so much!

  • @Ryeback101
    @Ryeback101 5 ปีที่แล้ว

    Excellent videos. I just went through the playlist and they explain the concepts really well. You sir are a hero!!

  • @shantanu991
    @shantanu991 6 ปีที่แล้ว

    I think, you made it so simple. I was looking for a proper explanation of this formula. Liked. Subscribed.

  • @hamadaparis3556
    @hamadaparis3556 4 ปีที่แล้ว

    this video demystified everything of the previous one, thank you so much

  • @samyakjain7300
    @samyakjain7300 5 ปีที่แล้ว

    A beautiful mathematical explanation of Gradient Descent! Way to go man...

  • @binilg
    @binilg 7 ปีที่แล้ว +2

    Man, you are super!! I had a hard time understanding the mathematics of gradient descent and you made it very easy. Thank u

  • @boemioofworld
    @boemioofworld 7 ปีที่แล้ว +9

    amazing explanation! One of the best explanation in the whole youtube IMO

    • @TheCodingTrain
      @TheCodingTrain  7 ปีที่แล้ว

      Glad to hear, thank you!

    • @idrisalhajiadamu7590
      @idrisalhajiadamu7590 2 ปีที่แล้ว

      @@TheCodingTrain thanks for the explanation. I couldn't find the next value where you explained batch gradient descend.

  • @anannyapal9939
    @anannyapal9939 6 ปีที่แล้ว +1

    I was literally so frustrated with these things meesing up my head.... thank you sir for helping me to survive....you are just fanstastic🙏🙏

  • @life_outdoor9349
    @life_outdoor9349 6 ปีที่แล้ว

    One of the best video tutorial I came across !!

  • @Kidkromechan
    @Kidkromechan 6 ปีที่แล้ว

    This is EXACTLY what i had been searching for the past week pfft.. Thank you Sir ^_^

  • @peterpanmjli
    @peterpanmjli 7 ปีที่แล้ว

    at 12:01 , the summation sign on the cost function just disappear (sum of error squared) ? What happened ?

  • @joesan3597
    @joesan3597 5 ปีที่แล้ว +2

    That was indeed very well explained! Thanks for the video! Could you please explain me why you treat the J(m, b) = ERROR^2 and then substitute the ERROR with mx + b -y when doing the partial derivative?

  • @user-xn4yu5rn9q
    @user-xn4yu5rn9q 5 ปีที่แล้ว

    At first I thought this is BS, now I’m so thankful

  • @michelaka6836
    @michelaka6836 7 ปีที่แล้ว

    Superb break down of this often miss-explained concept.!!!! A+

  • @awsomeguy563
    @awsomeguy563 7 ปีที่แล้ว +2

    Hi, im very confused. If you took the partial derivative of the cost function with respect to m, wouldnt you get the direction of steepest ascent. Shouldnt we negate that to get the direction of steepest descent? Shouldnt we minus the gradient from m????

    • @awsomeguy563
      @awsomeguy563 7 ปีที่แล้ว

      I found the answer! It all depends on how you define error, if you defined error to be predicted - target, then it makes sense to subtract delta m, but if you made error target - predicted, then it makes sense to add delta m, and this has to do with taking the partial derivative and then negating it. So Mr. Shiffman, it would be better that you mentioned this, as what you define error to be is very important. and helps with why you add delta m instead of subtract

    • @jetsmite
      @jetsmite 7 ปีที่แล้ว

      thank you i was having the same problem!

    • @jetsmite
      @jetsmite 7 ปีที่แล้ว

      where did you find the answer?

    • @awsomeguy563
      @awsomeguy563 7 ปีที่แล้ว

      jetsmite Haha, jetsmite, it bugged me so much, I was determined to find an answer. I asked on stackoverflow and on other youtube videos, nobody replied. One day I was in the bathroom and going through the derivative in my head, and I figures it out !

    • @jetsmite
      @jetsmite 7 ปีที่แล้ว

      awsomeguy563 could you expound on your answer? Im still trying to wrap my head around it? sorry about that LOL

  • @adacristina3763
    @adacristina3763 6 ปีที่แล้ว +3

    Amazing! The best explanation so far

  • @TheWeepingCorpse
    @TheWeepingCorpse 7 ปีที่แล้ว +12

    thank you for everything you do. I'm a c++ guy but your videos are very interesting.

    • @thecodersexperience
      @thecodersexperience 7 ปีที่แล้ว +2

      hey can u tell me best place to learn c++

    • @syedbaryalay5849
      @syedbaryalay5849 6 ปีที่แล้ว +1

      plus i would not jump into c++ as my first langauge, try learn an easy langauge and then start with c++. plus you need to be sure what are you gonna c++ for.

    • @CamdenBloke
      @CamdenBloke 6 ปีที่แล้ว

      Deitel and Deitel's book.

    • @douloureuxcrouton5780
      @douloureuxcrouton5780 6 ปีที่แล้ว +1

      gibson If his goal is to only learn C++, then learning C first is unnecessary. I would even argue it is a big mistake.

  • @adityarprasanna
    @adityarprasanna 6 ปีที่แล้ว +1

    It doesn't matter whether it's (guess - y) or (y - guess) in the cost function because it's being squared anyway right?

  • @8eck
    @8eck 4 ปีที่แล้ว

    So at 19:15, it's a formula to get the actual delta value of the cost/loss for one epoch?

  • @rupalagrawal3534
    @rupalagrawal3534 5 ปีที่แล้ว

    its a great video.A simple and easy language is used to explain every concept.Great work!!

  • @prachikhandelwal2259
    @prachikhandelwal2259 5 ปีที่แล้ว +1

    I wish I had a teacher like you. You are amazing sir I think it doesn't matter whatever are you studying but your teacher has the power to make the concept Easy or Difficult.
    And you are the one who makes everything extremely easy!
    and Yeah you are damn funny.

  • @harshgulati6189
    @harshgulati6189 2 ปีที่แล้ว

    Please try to complete this whole series with all basic ML algorithms

  • @dmill989
    @dmill989 4 ปีที่แล้ว

    Great explanation thank you! I kept seeing the chain rule, which I understand, but no one was explaining explicitly which chain of functions we are using it on and that the loss function is at the end of the chain.

  • @satyaNeelamraju7
    @satyaNeelamraju7 7 ปีที่แล้ว +1

    An excellent video.. The best video in the internet for Gradient Descent Algorithm. Thanku so much :) ... Keep posting like this

  • @mikelosei382
    @mikelosei382 3 ปีที่แล้ว

    very, very, very helpful!!! I'm in grade 12 and was researching how exactly calculus could be applied to com sci, and this was a life saver! I had no idea what I was doing before this XD Thankss

    • @TheCodingTrain
      @TheCodingTrain  3 ปีที่แล้ว +1

      So glad to hear thanks for the nice comment!

  • @ATXpert
    @ATXpert 4 ปีที่แล้ว +1

    i got stuck at the chain rule in 3blue1borwn, watched this, it took some thinking but i understand it roughly enough to use it

  • @michaelstopa1774
    @michaelstopa1774 4 ปีที่แล้ว

    Dude, you are completely mad. But in the most noble sense of this word;)! Fantastic way of explaining an actually quite complex piece of math. And it's very funny too;). Congrats and hats off. You're an excellent educator.

  • @pramodkhatiwada6189
    @pramodkhatiwada6189 6 ปีที่แล้ว

    It's awesome, i understand the concept of error minimizing and i jump over here to comment.

  • @lightningblade9347
    @lightningblade9347 6 ปีที่แล้ว

    Instant subscription, I adore the passion you have for what you do :) .

  • @georgyandreev7469
    @georgyandreev7469 4 ปีที่แล้ว

    Man that’s awesome. Being Russian I’ve understood every single thing. Keep up!

  • @kenhaley4
    @kenhaley4 7 ปีที่แล้ว

    Well, since I gave you a negative review on the calculus video, I feel I owe you one here.
    I thought that was great! The only thing you glossed over was the fact that the cost function is actually a summation of all the errors of each x value. But, since the derivative of a sum is simply the sum of the derivatives, putting the computation of m and b inside the for loop works fine. (Seeing that in your code at first actually bothered me, but now I see that it's no problem--it's exactly what's needed.) I found it fascinating how simple everything turns out after going through all the calculus. And I think that was the important point. Nice job.

    • @TheCodingTrain
      @TheCodingTrain  7 ปีที่แล้ว +1

      Thank you, I really appreciate it. I think there is still more room for improvement and, in a way, I'm just making these videos to help give myself the background for future videos. But I'm glad this one seems to be better received!

  • @snackbob100
    @snackbob100 4 ปีที่แล้ว

    this video is fantastic, you are a very talented teacher

  • @99chintu
    @99chintu 6 ปีที่แล้ว +1

    That is one of the best explanations I saw on the youtube..Thanks a lot..

  • @Raghad-mz8el
    @Raghad-mz8el 6 ปีที่แล้ว

    thank you for explaining this. and even better than what my professor did in multiple lectures.

  • @mathhack8647
    @mathhack8647 2 ปีที่แล้ว

    I lov this.
    " I tried again " Love this ,

  • @jebastinkoilraj1951
    @jebastinkoilraj1951 5 ปีที่แล้ว

    OHH man.. you're one hell of a teacher... Loved it

  • @divyagulakavarapu9614
    @divyagulakavarapu9614 5 ปีที่แล้ว

    Nice one. I actually had a doubt in GD, but watching your video I think I'm clear a bit.

  • @debasishhazra3222
    @debasishhazra3222 4 ปีที่แล้ว

    wonderful way of teaching and just fantastic video. Just Loved it Man...!!!

  • @annperera6352
    @annperera6352 3 ปีที่แล้ว

    Thank you Sir.This teaching gained you a subscriber

  • @linghadharshananparasu8596
    @linghadharshananparasu8596 5 ปีที่แล้ว +1

    I have a question, why do you add the error and m, and not subtract the error from m? After all the loss function is preditction - y, so shouldn't you subtract error from m?

    • @tomzhang3597
      @tomzhang3597 5 ปีที่แล้ว

      I don't know why he changed it to guess-y. But if you look in the code its y-guess, so if m is bigger, guess is bigger and error is smaller.

  • @dipakgaikwad3664
    @dipakgaikwad3664 ปีที่แล้ว

    You are awesome, atleast tried not skip the mathematics, not like most of them who just run away from mathematics

  • @danielkashkett7040
    @danielkashkett7040 4 ปีที่แล้ว

    This really helped me understand the MSE derivative. Great job!

  • @aGh-rw7dd
    @aGh-rw7dd 7 ปีที่แล้ว

    No worries I understood this lecture and appreciate it. I have studied Calculus 1 &2 in my high school

  • @AdityaSingh-lf7oe
    @AdityaSingh-lf7oe 4 ปีที่แล้ว +1

    Thank u so much!!! This is just what I needed... U rock!!

  • @crehenge2386
    @crehenge2386 7 ปีที่แล้ว +1

    it's interesting how different youtube channels become different classes^^ Khan gives you allt the calculus you could ever want if you're a beginner

  • @chandimaindatissa6562
    @chandimaindatissa6562 6 ปีที่แล้ว

    Simple and easy to understand. Thanks for sharing other important links.
    Well done!!

  • @kangjawab2740
    @kangjawab2740 6 ปีที่แล้ว

    Did you miss on the second of derivative function in your chain rule? It should be negative X (-x) and -1 respectively for dError/dM and dError/dB.

  • @Omar-kw5ui
    @Omar-kw5ui 5 ปีที่แล้ว

    Excellent explanation. Although I must point out that we travel in the direction of the negative of the gradient. So we multiply by -(learning rate)

    • @luthfiramadhan7591
      @luthfiramadhan7591 5 ปีที่แล้ว

      sir can u explain pls i dont gwt it why this guy + the gradient

  • @danieleps1926
    @danieleps1926 6 ปีที่แล้ว +1

    @The Coding Train Once you have the cost function, @4:24, how does one graph it continuously, because only by doing that and by deriving an explicit formula can one compute the gradient surely?

    • @x-lightsfs5681
      @x-lightsfs5681 6 ปีที่แล้ว

      Yeah, the error would be different every iteration (i think)

  • @cimmik
    @cimmik 5 ปีที่แล้ว

    Thank you so much. Finally I found an explanation that I could undestand. Good job, Daniel :D

  • @olicmoon
    @olicmoon 6 ปีที่แล้ว

    it's crystal clear even for a person like me lacking understanding of calculus

  • @philips5130
    @philips5130 3 ปีที่แล้ว

    I love your energy! Nice explanation btw

  • @realcygnus
    @realcygnus 7 ปีที่แล้ว

    superb ! as always. Knowledge of the more advanced concepts/techniques, especially higher level maths/abstractions etc. is primarily, precisely what separates the boys from the men / the script kiddies from software engineers etc. fear it NOT ! he plays the part so well mostly as a great teaching strategy/angle so as to reach as many as possible & to make them feel they aren't alone. Even if he really doesn't like it. I'd like to think its mostly an act anyway. & the Oscar goes to: Dan Shiffman ! He do knows his shiz though ! these vids have been priceless to me. thanks ! can't wait for the rest of Neural Networks

  • @rashariyad1821
    @rashariyad1821 3 ปีที่แล้ว

    Thank you, its very simple yet amazing explanation.

  • @bosepukur
    @bosepukur 7 ปีที่แล้ว +7

    very energetic presentaion ...loved it

  • @95030359503035
    @95030359503035 6 ปีที่แล้ว

    Could you explain the difference of the gradient with the mean square error and the mean absolute error?

  • @Suigeneris44
    @Suigeneris44 6 ปีที่แล้ว

    You're an amazing teacher! I wish I had a Math teacher like you!

  • @bap6491
    @bap6491 7 หลายเดือนก่อน

    Hi I am kinda late to the video but I wonder if you can help me? What are the Domain or Co domain of the hypothesis and cost function?

  • @furrane
    @furrane 7 ปีที่แล้ว

    Good job Dan. I study Math so this is ok for me, but I think this might be a little overkill for a coding enthousiast (or worst, a coding artist, haha) : while it is good to know the fundamentals, in practice, to get good results you'll usually want something a little "finer" than a linear regression to model your data and maybe use a more complex cost function, all of those dramatically increase the calculus you have to do to get the gradient descent working, meaning most people would use nice packages or already known formulas. This is just a suggestion but if it was me I'd focus more on what are the advantage of a given method, for example the squared error cost function "punishes" a lot more greater y/guess difference compared to say a simple absolute error cost function.
    Still a great video, please more =)

    • @TheCodingTrain
      @TheCodingTrain  7 ปีที่แล้ว +1

      This is very good feedback indeed. Yes, I agree that this topic is not really necessary. In a way, I'm really making these videos out of my own curiosity and I hope to be able to move more quickly to practical examples of how/why/etc to make stuff with these algorithms.

    • @furrane
      @furrane 7 ปีที่แล้ว

      Curiosity is good =) I enjoy maths but I'm impressed you pushed yourself to learn all this out of curiosity while your main field seems to be coding.
      If you feel like learning some more, you might want to take a look at Newton's method, it's used to find roots of a function (x is a root of f if f(x)=0) and it uses the same kind of iterations as gradient descent. Of course you can apply Newton's method to f' ( or df/dx ) to find minima or maxima of f.
      Anyway, have a great day and please keep doing those nice videos =)

  • @lxlyzd
    @lxlyzd 7 ปีที่แล้ว

    I saw somewhere that the loss function has the sum multiplied by 1/N ? Is it not important here ?

  • @yashpandey350
    @yashpandey350 4 ปีที่แล้ว +1

    After 5:30 your expression is like👏👏👏👏👏😂😂😂😂😂