Grant
Hey Grant, Love your video's from Khan and 3Blue1Brown!
Yo Khan Academy, thank you for making these videos. They are a real lifesaves at times ^w^
Awesome video! Thank you! And wow! It's 3Blue1Brown's voice doing this video!
Thank you so much, here I get much more infornation in one day than in university in a month)))
you are perfect, thanks for you videos, and you're fanny mood :)
So clear and helpful!
Awesome guy.... Mr Sanderson.
thank you so much ! so helpful
I love u Khan!! u save me today
Thank you sir, this video has given me a good idea
Thank you!
Wow, amazing. Thank you!
Thank you for the amazing explanation
고맙습니다
Can someone explain why the ideal learning rate for 2 or more dimensions in the gradient descent algorithm is the inverse of the Hessian (matrix of second partial derivatives)?
This guy explains it well: medium.com/@ranjeettate/learning-rate-in-gradient-descent-and-second-derivatives-632137dad3b5 . Intuitively, using the first derivative gives us the change in loss w.r.t x as a straight line; using second derivatives gives us information about the curvature of the loss function.
What does the Hessian matrix represent geometrically? In particular, what does the determinant of the Hessian matrix measure?
That's really a good question! Sadly I can't answer it now, but I'll use it as an inspiration to look into it, when I have the time. I think the best strategy to approach this problem is to calculate the determinant for some 2 or 3 dimensional functions and then play around with different values for x,y and z.
the moment I clicked on this link, oh this is the 3blue1brown guy!
So is this Hessian matrix is valid only for scalar valued functions right? If my intuition is correct then for a vector valued function of maybe 4 components, would there be 4 Hessian matrices?
what if the output of function f was a vector of 3 rows instead of a single expression. How would the hessian change?
awesome! :-)
i have a question
what kind of tools are you using when you work??
I really wanna get that blackboard tool :-) thx in advance
Is this the 3blue1brown guy?
Do you guys know which lecture/series/playlist is this video from? Please let me know! Thanks!
how do i know which channel of khan academy is for this video?
You can find this video on the Khan Academy website by using the search bar at the top of the screen and typing in "hessian."
Here is the link:
www.khanacademy.org/math/multivariable-calculus/applications-of-multivariable-derivatives/quadratic-approximations/v/the-hessian-matrix
Is there a vector form of the multivariable Taylor series?
I'm pretty sure that's the taylor expansion using the jacobian for derivatives.
interesting
Good day, I was wondering whether you know any python library that has implemented second order gradient descent with hessian error matrix. If you can point me to the right direction, it would be very helpful. Thanks in advance, Kind regards
Shantanu
Sal, if I have 1 equation and 6 independent variables, my partial first derivatives is a vector with 6 terms. If I follow, the Hessian will be a 6x6 matrix. Is that correct? Thanks!!! I contribute to you as your program and platform makes an amazing contribution!
when can fxy!=fyx ?
When will the diagonal not be symmetric?
@@ethanbooth1174 Sure of that ? I mean, for the second mixed derivatives to be different, they have to exist.
What if the function is a matrix itself? The Hessian matrix will be a matrix of matrices?
is he 3 blue 1 brown ???
i think in place of hessian you actually have mentioned hessian transpose!
DUDE and why don't you tell me how to find extrema points with this !!!
Gamer Sparta You find the eigenvalues of this matrix after solving the differential equation that optimizes. Find out if this matrix is positive definite, negative definite or semi definite.
hey, you are 3blue1brown?
Genius.
Ah, so this is where the formula for the discrimination comes from.We can see that taking the determinant of the Hessian gives the formula for the discriminant.I know it works for R^2. Will verify for R^3 and R^n as an exercise.Thanks!
fitz?
Is this the same guy as 3 blue 1 brown?
0:00 - 0:06 sorry what? Don't think I've ever been confused so quickly in a tutorial.
Is this 3blue1brown as the lecturer?
And the rest of the video?
👌👌👌👌
The video is in fundraiser but the video is from 3b1b hehh? 🧐🧐
When are Fxy and Fyx not equal?
@@matakos22 Sorry, I don't know what C2 means ... do you have an example please ?
@@joluju2375 Continuously differentiable functions are sometimes said to be of class C1. A function is of class C2 if the first and second derivative of the function both exist and are continuous. More generally, a function is said to be of class Ck if the first k derivatives f′(x), f′′(x), ..., f (k)(x) all exist and are continuous. If derivatives f (n) exist for all positive integers n, the function is smooth or equivalently, of class C∞.
@@matakos22 Thanks. So, for Fxy and Fyx not to be equal, they have to exist. Then, if F is not C2 and Fxy and Fyx exist, it means that Fxy or Fyx is not continuous. Right ?
Jay Shree Ram
Aren't you the guy of 3 Blue 1 Brown?
Dude I’m in 8th grade doing calc 1 and I already understand this.
Grant???
If you differntiate x first, then y, shouldnt it be "dxdy" ? why do you keep putting it backwards
No, the way he does it is notationally correct. Oh course, you are free to write things the way you like but he is following the convention.
Well you should think of it this way: d/dx (df/dy), so you take df/dy and then differentiate it with respect to x, so the video is correct. In other words we start with the partial derivative with respect to y and then differentiate it with respect to x.
Came for Neo and Morpheus, left disappointed.
First
la matrice de la hess (la misère) xD
I thought i clicked 3b1b's video
Aye 3B1B
if anyone is wondering why the mixed derivates are the same: it's Schwarz's theorem.