I loved this discussion! The combination of Alex knowing everything in full mathematical generality and Yannic trying to bring it down to the”real world” really helped me! I’m new at this subject, it really helps to walk through a toy problem such as temperature in a room using a simple basis and describing the vectors formed etc. so that it feels less nebulous to begin with. Granted, I’m an engineer so what’s best for me is first show me a simplified version and how it works concretely, THEN abstractify it to death to make it maximally useful. Thanks to all three of you!!! It amazes me that such great content is just “out there” to be found!
3 ปีที่แล้ว +4
I love your channel! Though I have to admit that I felt a little lost with all the terminology being thrown around when I first watched this video in particular. I decided to delve deeper into kernels and after intensive research, I have created a 6 hours long playlist on Kernel Methods to summarize my current understanding. If anyone wants to have a crash course on kernels in particular, I'd be delighted to welcome you in my comment section. After these countless hours of self study, I can now follow the conversation fully which is such a nice feeling of accomplishment. Thank you for inspiring me to research this topic in-depth!
Very helpful video, happy it exists! Perhaps the format could have allowed for some slides here and there, since Alex Stenlake has prepared an explanation in advance. Just to avoid him gesticulating the visualizations. And also verbalizing mathematical examples that are easy to understand when written, harder to follow when just spoken out loud. 😊
Around 39 minutes one bit that is missing is that an inner product of two functions is the integral of the functions multiplied together over the domain of their arguments.
I really wanted to appreciate the efforts but probably only people with already very good understanding about kernels can handle all these verbal discussions with abstract and repeatedly used words. I see that Yannic tried to make it clearer by asking some clarifying questions, though, which helped a little bit.
Just a naive viewpoint/question here, yannic in his video about linformer mentioned about the JL theorem which multiples a high dimensional data distribution with fixed gaussian distribution to lower dimensions while preserving the distance between data point constant. If kernels are also a distance similarity measure, which also kind of projects data from lower dimension to a certain higher dimension ( rewatching the video again i am at 17 min currently ) so pairwise distance measures between data points seems to be a sorta accurate representation for any distribution and any projecting from higher to lower or vice versa dimension must be focused on preserving the distance measure
I love this podcast. Really such a cool idea. I just wanna give some tips that might make it even more better, at least visually. It kind of really annoys to see the bad quality of the people talking. I think it would be so much cooler if everyone would record his camera and audio and afterwards send it to Tim who cuts it together in a way you have it now , where every person is visible at any time, just in way better quality. That way there are also way more options to make the design of the podcast cooler. For example you could put a nice layout over it or something. Also I feel like sometimes it would come in so handy if you would bring some pictures on the screen. A bit like Tim already did where he opened up the papers. It would look so much more professionell to the viewer and Im sure others would like it too. Anyways, I love the show
Amazing stuff! Also would be cool if you could talk about dimensionality reduction methods to solve the memory inefficiencies of kernel methods with large datasets.
Hawkz1600, it seems that the biggest breakthrough here to fix memory issues is the usage of "random features" to approximate general kernels by random linear kernels. See the paper "Random features for large-scale kernel machines. "
This would have been such a nice presentation if Tim didn't distract the flow of the conversation. Yannic tried a few times to recover the underlying goal of the Alex's talk, but failed. I wish this could be redone with the Alex talk on underlying concept and its application to the Yannic's room temperature model as a specific example.
I'd love to know anyone's thoughts on the usefulness/utility of 1) Random Fourier Features (a trick to approximate kernels by certain linear kernels, and thus speed computations up. ) 2) Reproducing kernel Banach spaces (doing kernel methods in a Banach space that promotes sparsity more than doing kernel methods in a Hilbert space setting would, kind of like Lasso regression vs Ridge regression. )
Your channel has way to few subscribers. Such good content, im not even a machine learning engineer, just a programer who learn this all stuff at the moment.
Why not having a board and writing on it. This make it more interesting. More importantly having fun examples can give audience much better understanding.
Talking about highly advanced mathematics without notational & visual support is highly silly. There is no one who can understand the english who cannot understand visuals, and many who could only understand the visuals.
I love how Yannic takes a step back and explains things using his intuition. very helpful!
Everything Yanic speaks is gold...I understand instantly
I know right 😂
I know! He always asks what I want to know, it’s kinda spooky how good of a communicator he is!
Without Yanic, I can't understand a word
The contribution of all 3 of you was significantly more than the sum of the parts. Very enjoyable, thanks.
best explanation of kernels I've encountered so far, thanks!
I loved this discussion! The combination of Alex knowing everything in full mathematical generality and Yannic trying to bring it down to the”real world” really helped me! I’m new at this subject, it really helps to walk through a toy problem such as temperature in a room using a simple basis and describing the vectors formed etc. so that it feels less nebulous to begin with. Granted, I’m an engineer so what’s best for me is first show me a simplified version and how it works concretely, THEN abstractify it to death to make it maximally useful. Thanks to all three of you!!! It amazes me that such great content is just “out there” to be found!
I love your channel! Though I have to admit that I felt a little lost with all the terminology being thrown around when I first watched this video in particular.
I decided to delve deeper into kernels and after intensive research, I have created a 6 hours long playlist on Kernel Methods to summarize my current understanding. If anyone wants to have a crash course on kernels in particular, I'd be delighted to welcome you in my comment section.
After these countless hours of self study, I can now follow the conversation fully which is such a nice feeling of accomplishment. Thank you for inspiring me to research this topic in-depth!
I loved this conversation with Alex! We already recorded 2 more casual conversions, we will upload them in the coming days
Very helpful video, happy it exists!
Perhaps the format could have allowed for some slides here and there, since Alex Stenlake has prepared an explanation in advance. Just to avoid him gesticulating the visualizations. And also verbalizing mathematical examples that are easy to understand when written, harder to follow when just spoken out loud. 😊
Around 39 minutes one bit that is missing is that an inner product of two functions is the integral of the functions multiplied together over the domain of their arguments.
Thank you, that makes more sense now. But the result, when evaluated, should still be a scalar, right?
@@oblomist in this case yes, not sure if that's true in general.
Wow! Such good and informative video 😃
Thank-you Abby!
love your videos .
I really wanted to appreciate the efforts but probably only people with already very good understanding about kernels can handle all these verbal discussions with abstract and repeatedly used words. I see that Yannic tried to make it clearer by asking some clarifying questions, though, which helped a little bit.
Just a naive viewpoint/question here, yannic in his video about linformer mentioned about the JL theorem which multiples a high dimensional data distribution with fixed gaussian distribution to lower dimensions while preserving the distance between data point constant. If kernels are also a distance similarity measure, which also kind of projects data from lower dimension to a certain higher dimension ( rewatching the video again i am at 17 min currently ) so pairwise distance measures between data points seems to be a sorta accurate representation for any distribution and any projecting from higher to lower or vice versa dimension must be focused on preserving the distance measure
Top quality content👌👌
Great talk guys! Reminds me why I got into ML in the first place, and reevaluate Bishops book 😅
Serge, what is the title of this book? I'd like to find myself a copy.
@@bradleypliam110 Pattern Recognition and Machine Learning
Great book!
@@SergeTheGod Thank you for the leg up!!
11:50 where the kernel talk started
I love this podcast. Really such a cool idea. I just wanna give some tips that might make it even more better, at least visually. It kind of really annoys to see the bad quality of the people talking. I think it would be so much cooler if everyone would record his camera and audio and afterwards send it to Tim who cuts it together in a way you have it now , where every person is visible at any time, just in way better quality. That way there are also way more options to make the design of the podcast cooler. For example you could put a nice layout over it or something.
Also I feel like sometimes it would come in so handy if you would bring some pictures on the screen. A bit like Tim already did where he opened up the papers. It would look so much more professionell to the viewer and Im sure others would like it too.
Anyways, I love the show
Tim's breakfast bar/kitchen island arrangement is impressive, but the tin of Coffee Mate hurts my soul.
Lol!!! But what are you saying here? 1) You don't like the design on the tin 2) you don't like the manifold of the tin 3) you don't like coffee mate 😂
Amazing stuff! Also would be cool if you could talk about dimensionality reduction methods to solve the memory inefficiencies of kernel methods with large datasets.
Hawkz1600, it seems that the biggest breakthrough here to fix memory issues is the usage of "random features" to approximate general kernels by random linear kernels. See the paper "Random features for large-scale kernel machines. "
This would have been such a nice presentation if Tim didn't distract the flow of the conversation. Yannic tried a few times to recover the underlying goal of the Alex's talk, but failed. I wish this could be redone with the Alex talk on underlying concept and its application to the Yannic's room temperature model as a specific example.
Sorry! Feedback taken on board
I'd love to know anyone's thoughts on the usefulness/utility of
1) Random Fourier Features (a trick to approximate kernels by certain linear kernels, and thus speed computations up. )
2) Reproducing kernel Banach spaces (doing kernel methods in a Banach space that promotes sparsity more than doing kernel methods in a Hilbert space setting would, kind of like Lasso regression vs Ridge regression. )
Your channel has way to few subscribers. Such good content, im not even a machine learning engineer, just a programer who learn this all stuff at the moment.
"infinite dimensional, or high dimensional, or don't-wanna-compute-able" haha!
"and that's because least-squares is a horrible, blurry loss function" =)
Solid
Why not having a board and writing on it. This make it more interesting. More importantly having fun examples can give audience much better understanding.
Talking about highly advanced mathematics without notational & visual support is highly silly. There is no one who can understand the english who cannot understand visuals, and many who could only understand the visuals.
Lose the shades.