- 83
- 88 217
Mediterranean Machine Learning (M2L) summer school
เข้าร่วมเมื่อ 22 ธ.ค. 2020
Support the ML community in the Mediterranean countries, and encourage diversity in ML.
วีดีโอ
[M2L 2024] A Brief Tutorial on Robot Learning - Georgia Chalvatzaki
มุมมอง 873 หลายเดือนก่อน
[M2L 2024] A Brief Tutorial on Robot Learning - Georgia Chalvatzaki
[M2L 2024] Graph Neural Networks - Helena Andres
มุมมอง 1853 หลายเดือนก่อน
[M2L 2024] Graph Neural Networks - Helena Andres
[M2L 2024] (Generative) AI for Software Engineering - Leonardo Mariani
มุมมอง 1303 หลายเดือนก่อน
[M2L 2024] (Generative) AI for Software Engineering - Leonardo Mariani
[M2L 2024] Code Generation - Kate Baumli
มุมมอง 1173 หลายเดือนก่อน
[M2L 2024] Code Generation - Kate Baumli
[M2L 2024] RLHF - Daniele Calandriello
มุมมอง 1363 หลายเดือนก่อน
[M2L 2024] RLHF - Daniele Calandriello
[M2L 2024] Planning and Reasoning - Theophane Weber
มุมมอง 1223 หลายเดือนก่อน
[M2L 2024] Planning and Reasoning - Theophane Weber
[M2L 20204] A Friendly Tasting of Reinforcement Learning: Foundations and Algorithms- Claire Vernade
มุมมอง 1483 หลายเดือนก่อน
[M2L 20204] A Friendly Tasting of Reinforcement Learning: Foundations and Algorithms- Claire Vernade
[M2L 2024] Mixture of Experts - Diego de Las Casas
มุมมอง 1503 หลายเดือนก่อน
[M2L 2024] Mixture of Experts - Diego de Las Casas
[M2L 2024] Large Language Models - Thomas Mesnard
มุมมอง 953 หลายเดือนก่อน
[M2L 2024] Large Language Models - Thomas Mesnard
[M2L 2024] Machine Learning for Theorem Proving - Emily First
มุมมอง 1403 หลายเดือนก่อน
[M2L 2024] Machine Learning for Theorem Proving - Emily First
[M2L 2024] Introduction to Simulation Free Generative Models - Joey Bose
มุมมอง 4353 หลายเดือนก่อน
[M2L 2024] Introduction to Simulation Free Generative Models - Joey Bose
[M2L 2024] Visual Generative Models as Queryable World Models - Adriana Romero Soriano
มุมมอง 693 หลายเดือนก่อน
[M2L 2024] Visual Generative Models as Queryable World Models - Adriana Romero Soriano
[M2L 2024] Generative Models - Marta Garnelo
มุมมอง 2983 หลายเดือนก่อน
[M2L 2024] Generative Models - Marta Garnelo
[M2L 2024] Quantum Machine Learning - Armando Bellante
มุมมอง 1443 หลายเดือนก่อน
[M2L 2024] Quantum Machine Learning - Armando Bellante
[M2L 2024] Transformers - Lucas Beyer
มุมมอง 2.1K3 หลายเดือนก่อน
[M2L 2024] Transformers - Lucas Beyer
[M2L 2024] Optimisation in Deep Learning - Mihaela Rosca
มุมมอง 1733 หลายเดือนก่อน
[M2L 2024] Optimisation in Deep Learning - Mihaela Rosca
[M2L 2024] Introduction to Deep Learning & Vision - Dilara Gokay
มุมมอง 4953 หลายเดือนก่อน
[M2L 2024] Introduction to Deep Learning & Vision - Dilara Gokay
2023 1.1 Christos Papadimitriou - AI and the Brain
มุมมอง 1.2Kปีที่แล้ว
2023 1.1 Christos Papadimitriou - AI and the Brain
2023 2.3 Creativity and ML - Konstantina Palla
มุมมอง 192ปีที่แล้ว
2023 2.3 Creativity and ML - Konstantina Palla
2023 5.3 How to train Neural Networks effectively - Sander Dieleman
มุมมอง 1.7Kปีที่แล้ว
2023 5.3 How to train Neural Networks effectively - Sander Dieleman
2023 6.1 The practice of Research Engineering - Fabio Viola
มุมมอง 260ปีที่แล้ว
2023 6.1 The practice of Research Engineering - Fabio Viola
2023 2.1 VAEs and GANs - Xenia Miscouridou
มุมมอง 332ปีที่แล้ว
2023 2.1 VAEs and GANs - Xenia Miscouridou
2023 1.3 Transformers - Sahand Sharifzadeh
มุมมอง 1.2Kปีที่แล้ว
2023 1.3 Transformers - Sahand Sharifzadeh
2023 3.1 Personalised Longitudinal NLP - Maria Liakata
มุมมอง 114ปีที่แล้ว
2023 3.1 Personalised Longitudinal NLP - Maria Liakata
2023 3.3 Predicting the Past - Yannis Assael
มุมมอง 121ปีที่แล้ว
2023 3.3 Predicting the Past - Yannis Assael
When mentioning KV-cache, Lucas incorrectly highlights the sequence length as the embedded vector length, when the latter is the d_model, and sequence length is the context window
it is at 39:35 when he drags the point horizontally instead of vertically. this is a very minimal error and I have no doubt Lucas did this by accident, but I am just highlighting in case somebody else notices this and isn't sure. If I am wrong about this someone please tell me!
at 37:00, why are the special tokens different for the two sequences?
as in <bos> vs. <eos>
❤❤
Fourier transform is mentioned. How does one learn the applications of that in the context of MLL?
please make it 2k or 4k
beyer 😍😍😍
tnx for sharing
As an AI expert and teacher, I love the way you present the threats and challenges of AI in a way that has room for hope. I ordered your book and will likely use it in my Machine learning course to let computer science engineering students reflect on the future of AI. If ever you're in Belgium, I'd love to invite you for a guest lecture at Ghent University, from which many students (and professors) would surely benefit!
RL should thank Schopenhauer for this idea. Intellect is just servant to the Will i.e. will to live and procreate.
Artificial Intelligence is a Discovery not an invention. I was showing someone the AI bots that learned to play soccer. They stated "But when I play soccer , I know i am playing soccer." and so my response was. "Well you think that, because your Brain gives you a reward when you do." Meaning is a Human illusion.
I believe the figures around 13:00 and on the next slide are misleading. To my understanding the attention weights softmax(QK^T) would always be a square matrix TxT matrix with T the sequence length. Also the dimensionality of the keys and queries needs to be the same, otherwise the inner product in softmax(QK^T) would have a shape mismatch.
Thank you i loved it. And now there is a new book from him available on the topic: “Deep Learning”.
I didn’t expect fish to have more synapses then rats.
Batchnorm can be statefull, but most optimizers should be able to fuse the affine transformation into the convolution layer before it. In that case BN has no computation cost at runtime
nice presentation for beginners.
exciting to find this on yt!
How can I add 'Self-Attention mechanism' to Vector Quantized Diffusion Model for Text-to-Image Synthesis?
Really good lecture, thanks for uploading!
Thanks
Great course. Great human being.
Very very good overview, thanks for sharing!
This was the best overview of transformers I’ve seen to date. Great talk!
th-cam.com/video/-hcxTS5AXW0/w-d-xo.html q_{\phi} instead of q_{\theta}?
Yes it should be the case
Good presentation 🎉
Poor audio quality.
þrðmð§m 🌸
Essentially, reward is enough is just saying optimizing an objective is enough. That is definitely plausible, provided there is sufficient expressivity of the system to learn complex behaviour needed for the objective in the environment. The only issue I have is that the reward signal may be sparse, or stochastic, which would require many runs of the environment to learn it. How then, are humans/animals able to learn from limited trials and not kill ourselves in the process. Imagine if humans learn how to drive cars like how a DQN / A3C agent learns. There would be a million car crashes before the car could get out of the car park. Hence, reward is good and important - who drives the reward (external or internal) is a good point to think about. Could some behaviours also be innate?
Excellent webinar
Has cognitive 'science' inhibited the development of AI?
Thank you for sharing, insightful!
Vim após assistir o podcast com Serjão é muito bom saber que você está avançando no projeto boa sorte!
it is possible to update full course ty!
This kind of coverage of a published paper, especially this paper, is extremely helpful. Why this has only 3.6k views I cannot say. This one is big.
Also, seeing Sutton on there -- that's fantastic.
Thanks for sharing.
Thanks for posting this professor Cartis, I look forward to watching these lectures.
12:19 If a robot maximising only cleanliness is around children, the children are in trouble
Haha, so constrained rules are needed.
I find this hypotheses very plausible, but there needs to be a better effort at defining the working terms, such as intelligence. Otherwise, there is really no argument.
9:40 I didn't know pain is a reward :D
It is; it's a *negative* reward.
He literally told "reward is enough, but actually not enough".
thank you Andrea and organizers. Insights exposed are top level, please consider that the loop mentioned also scale up as a meta-analogy and is really amazing how one promising tool for understand the whole complexity of the brain is clearly an emulation of its proper function.. to investigate itself, ourselves. A friend of mine is a "language key" believer, and this presentation points that.
Is it possible to provide a link for the slides of this talk? Thanks a lot.
Sorry ' Do you use Matlab or python as your framework to apply your methodology ?
I make use of python