Interesting you'd say that, there is so much to learn even from asian schools like in India for example, but some accents can be harse when you watch the videos. Whats interesting about what you said is I think Andrew Ng also speaks eloquently and his lectures are worthy of a listen, and its the same topic area.
Hey Lex, can´t thank you enough for splitting up the day-long streams! Much easier to consume -- as I wanted to download to enjoy it on mobile! I came across Andrej Karpathy´s Deep Learning for Computer Vision yesterday. I´ve been trying to really understand CNNs and the Deep Learning paradigms technically for some time now and sucked up everything I could since July. Andrej´s lecture is the very best I found to get up to speed on most important details in the least possible time. Cheers G.
Yest i watched it 2 days before i watched this one as well. Andrej Karpathy's lecture is absolutely amazing. He covers so much in such a short time with such detail.
27:02 How are the gradients computed? 1) Gradients of the loss with respect to the activation 2) Gradients of the mean/sum of activations with respect to the input image
Ok I disagree with how he answered the woman's question. There is a way to choose a MINIMAL amount of layers, based on the complexity of the data. It depends on convexity vs non-convexity in the data space. non-convexity requires a minimum of 2 hidden layers to represent. This is why before you dig into deep learning and all the modern stuff a good basis in Universal Approximation Theory and NN's as Universal Approximators is critical!
Extra comment, so the important link is visible without expanding: I´m looking forward to dig into the extended version, namely the CS231n Winter 2016 Lecture series referenced in the video. (at th-cam.com/video/g-PvXUjD6qg/w-d-xo.html )
Anybody just love the way Karpathy speaks? Makes me want to listen to him with intent
same, he just gets to the meat of the stuff right away
Interesting you'd say that, there is so much to learn even from asian schools like in India for example, but some accents can be harse when you watch the videos. Whats interesting about what you said is I think Andrew Ng also speaks eloquently and his lectures are worthy of a listen, and its the same topic area.
Agreed. Idk if it is the nervous pause or else but he is so interesting when he explains stuff.
omg! He has done a great job in explaining this entirely new field of research in an hour!
Hey Lex,
can´t thank you enough for splitting up the day-long streams! Much easier to consume -- as I wanted to download to enjoy it on mobile!
I came across Andrej Karpathy´s Deep Learning for Computer Vision yesterday.
I´ve been trying to really understand CNNs and the Deep Learning paradigms technically for some time now and sucked up everything I could since July.
Andrej´s lecture is the very best I found to get up to speed on most important details in the least possible time.
Cheers
G.
Yest i watched it 2 days before i watched this one as well. Andrej Karpathy's lecture is absolutely amazing. He covers so much in such a short time with such detail.
How did this went under my radar for 7+ years
Hi from Brazil!! Thank you for sharing this. It is a really good material for researchers and who is initiating in this area.
Great video. I came here after completing 4 video sessions by Lex (MIT 6:S094). Thanks for compling these videos.
I love listening to things I cannot understand
Keep listening, you will eventually understand
makes u feel smart uh? cause i felt it that way too 😂
Don't say that, give it time.
Same here.....😂😂🤦♂️
For real bro . For real... 🤣🤣🤣🤣
that deepvis demo was kinda magical and mindblowing..
This is a great talk, really helped with my university project
The first terminators on the planet would sound exactly like Andrej Karpathy!... love to see him say 'I'll be back' :)
= = I have a feeling that I accelerate the video.....since he speaks so fast
@@Penaming 1.25 speed is even better
He just speaks very staccato, I think he speaks normal speed
Thank you lex sir for posting this
27:02 How are the gradients computed?
1) Gradients of the loss with respect to the activation
2) Gradients of the mean/sum of activations with respect to the input image
I think it compute the lost in the last layer and do the back propagation to compute the gradient of each layer so as to optimize.
That is choice number 1.
This People, They did not understand anything but they liked him.
So amazing...
Great lecture - thanks! :)
Very recommendable! Good overview!
Amazing
Good explanation
slides: docs.google.com/presentation/d/1Q1CmVVnjVJM_9CDk3B8Y6MWCavZOtiKmOLQ0XB7s9Vg/edit#slide=id.p
David Lee Thanks for sharing deck :)
Thanks! :)
Pro-Tip: 0.75x speed. You're welcome :)
he is one of the wingmakers
Ok I disagree with how he answered the woman's question. There is a way to choose a MINIMAL amount of layers, based on the complexity of the data. It depends on convexity vs non-convexity in the data space. non-convexity requires a minimum of 2 hidden layers to represent. This is why before you dig into deep learning and all the modern stuff a good basis in Universal Approximation Theory and NN's as Universal Approximators is critical!
the fun starts @28:14
This why youtube has playback speed settings.
🤓
thanks frid
does anyone know why lex hasn't had karpathy on the podcast yet?
I ask myself this question as well. Just wondering!
so.. know he had
Does someone have 2016, 2017, 2018, and 2019 winning architectures list?
recommend to watch it on x0.75 ))
can anyone load the slides of this presentation
Cool, thanks
Cell phone distributed training is a radical idea.🤔🤔🤔
Great
Extra comment, so the important link is visible without expanding:
I´m looking forward to dig into the extended version, namely the CS231n Winter 2016 Lecture series referenced in the video. (at th-cam.com/video/g-PvXUjD6qg/w-d-xo.html )
At 1:19:40. Isn't that what Google is trying to do right now?
cs231n
slides
Low resolution.
даня кода