Watch Full AI Series: th-cam.com/video/YulgDAaHBKw/w-d-xo.html NO MUSIC version: th-cam.com/video/DFDOyMZw9Q4/w-d-xo.html Sponsored by Brilliant | Use brilliant.org/artoftheproblem for 30-day free trial and 20% discount
Love your videos! Just a little suggestion. Background music is good, but too loud. It should never cover your speaking, like now. I suggest 15 - 20% less music volume, and you are good!
I love this video. I remember watching your videos like 10 years ago on Khan Academy about compression, entropy, Claude Shannon, etc. All timeless. I have always loved this style of documentaries. We need to protect you at all costs.
I love your video aesthetics, how you blend retro video clips with your explanations. I think you'd really enjoy retro-futuristic concepts and games like Bioshock and Fallout.
I like this kind of history lesson instead of learning advance topic directly. it gives us an idea how things were and explain things are the way there are
It is a great explanation of how current AI models reason. I liked the video a lot! 1. Simulation of future states. 2. LLMs that can give kind-of accurate answers with step by step reasoning. 3. RL approach that makes LLMs to give multiple answers, then evaluate them to select the best one. (Required more time) It would be nice to see wether a model that wasn't trained on the internet data, could learn how to reason by interacting with an LLM, and practicing on its dreams, but maybe we'll see that in the future. For the awesome review, history explanation and divulgation: Thanks! 🎉
I am not a subscriber? I remember watching this channel about 12 years ago. I found it again, and it keeps creating art out of problem solving. Good job!
@@ArtOfTheProblem cool i'm subscribed... i always wondered exactly how the "reasoning worked". What I remember from your video is that like with chess instead of trying all the games, it randomly picks 100 of them. So the same with the reasoning..
THANK YOU for publishing a no-music version of this video (see pinned comment by ArtOfTheProblem). It is such a clear and informative video that I hated to see it loose views due to the competing sound track. I'm going to watch it again right now to see if I missed anything the first time around. Thanks again for being so responsive to your followers.
I'm loving almost everything about this: the editing, the subject matter, the music. But as one other commenter alluded to; the audio mixing really falls short, especially near the end. Please consider making the background and effects less prominent going forward, it really sucks having to strain just to hear your voice, which is what we're here for! Subscribed ❤
Albert Einstein said: "If you can't explain something in a simple way so anybody can understand it you don't fully understand it yourself". Perhaps you are one of the few LLM experts we have!
Having read a bit about the AI safety arguements, learning about these arguably incredible developments into artificial minds is now accompanied by a sense of dread as well as the sense of awe.
Amazing explanations, visuals, and historical context! IIRC MuZero trained the policy and value networks (used to rollout the MCTS tree) also on the output of the MCTS tree. This seems super useful because search can be used to improve the the training of the networks (not just the results at inference time). I wonder if this also works for CoT/ToT in LLMS where the pretraining could include ToT to boost training performance?
yes it did, and yes it seems to help. Look at inference time training, just a few days ago a group got a new record on the ARC test doing this kind of thing (i haven't had time to go deep). x.com/akyurekekin/status/1855680785715478546
anothr great video. Understanding the "world model" and the algorithm that makes the decisions in it was very expansive. Also adding the self training/emulation of dreams is a powerful analogy to the human. seeing how thinking longer, blended in with intuition to make better chains of thoughts is also fantastic. Every time I reflect on machine learning, I learn more about myself. Which kind of makes you think its more sentient if it reminds me of myself? or the best emulator ever!
Great video as always, cant wait for the next ones! Top research quality. I think world models deserve more focus rather than llms, which are probably a dead end to true understanding of the real world. Yann lecun has very interesting ideas about these, in his JEPA and V-JEPA architectures and some of his lectures. I also think neuroscience can provide incredibly interesting and valuable insight into ml architectures as why not take ideas from a model undergone hundreds of millions of years of optimization for the same very abilities we are trying to model. Maybe memory is an interesting pathway (perhaps for a video), both working memory and long term (episodic, semantic)... Anyways, just some of the ideas I've been thinking about recently.
@@ArtOfTheProblem I for one was looking up some "how does AI work" stuff yesterday and some of your vids came up a couple of times, I watched multiple authors with their own unique takes (3Blue1Brown and Nottingham Uni's Computerphile also good channels). This video made me follow tho. I think you earned it :3
I would love to see a video like this on training LLMs and AI in general on morality. How to stop a decision tree that results in a positive outcome but arrives at it through immoral choices or actions.
That would be difficult but worth trying... . First it has to be agreed upon what's moral, you know starting out from the absolute that states everything is relative... .
I would love to see a video about the ethics of machine learning models and especially LLMs. There is a healthy body of literature out there to draw from about issues like intellectual property and copyright, enabling and obscuring bias, impact on marginalized communities, the resources used by model training and computation, etc
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! th-cam.com/video/PvDaPeQjxOE/w-d-xo.html
Very well done. One minor suggestion, the little sound effects could be less disruptive- maybe lower volume and fewer in number and duration? There is so much great explanatory detail, but I think some will find the extent of sound effects used a bit disruptive to the listening/learning process.
Great video. Discussing the definition of reasoning will probably be a moot point if we can all do the same things ....The difference between us and machines is that we aren't mere machines is that we have life and choices. The machine can be turned on and off, and only does as much as it is programmed, or in this case, trained, to do-which is what limits it to achieve something closer to AGI: agi needs a robot to sense the world, to understand the world. However, that's limited to the physical world. It won't understand our emotional world because it doesn't feel emotions, and it doesn't understand morality because it doesn't have a sense of morality as we do, we have to teach it that; and it isn't self motivated, so it's not responsible for anything-we are responsible for the goals we direct it to do. We have self-motivation and the free will to act on our motivations.
love these thoughts...on sense, did you see my video here th-cam.com/video/Dov68JsIC4g/w-d-xo.html (physical symbols...) on emotions, i've thought of this as 'learning signals' (did you see this: th-cam.com/video/5EcQ1IcEMFQ/w-d-xo.html) on free will...i wonder how it differs, seems like the boundary to explore further - does it matter where the goal came from?
I remember all of these developments and the never seemed like a big deal at the time. At the time It felt like winning at only chess meant we still had a long way to go.
Hi! I just finished the series. Great as usual! I would like to read up a bit more on the ML algorithms and networks that are behind these LLMs. I saw in this series that you kept referring to some papers, also highlighting parts of them with the marker. If you have a bibliography or a comprehensive list of these documents, would you mind sharing it please? Thanks!
awesome! I have one more summary video coming soon. i try to show all the papers so it's easy to look up, let me know if there is something you are looking for specifically
@@ArtOfTheProblem Will wait for the video, thanks! I am planning to understand better the distinction between classic NNs (such as the ones used for identifying digits) and transformers, since I will start a project in which I will have to fine tune some LLMs to try to demonstrate whether or not they are capable of identifying logical fallacies in political debates.
One tweak that would help this video perform better is to decrease the relative volume of the background music, especially at the end right before the ad. But it may be too late for that on this one, idk how TH-cam works.
Wow... I just rediscovered this channel. I remember watching your RSA and cryptography series around the time I purchased my first bitcoin and now I'm an Ethereum developer. This video was good, I don't have much input right now but I'm glad I found your channel again.
@@ArtOfTheProblem I'm working on community whistleblowing. I can never post links in youtube comments but if you google the title below you can find my paper: TandaPay Whistleblowing Communities: Shifting Workplace Culture Towards Zero-Tolerance Sexual Harassment Policies
@artoftheproblem I agree! I love the music. But maybe if you just lower its volume compared to the narration, then you might appeal to more people without losing those of us who like the music (but not necessarily its intensity). I think ones who complain might just be easily distracted by the soundtrack’s loudness rather than hate the music choices.
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! th-cam.com/video/PvDaPeQjxOE/w-d-xo.html
@@Farligefinn You know what, I just wrote a paragraph disagreeing with you but I reread the initial question and deleted it. Rereading and reinterpreting "Does it think actually?", I actually see what you're saying. A clearer word would have been "reason." "Think" can have a few different interpretations and I was contributing it towards consciousness. But whether AI "reasons" is a very different question entirely and I believe what him and you mean. Does it go through a sequence of logical steps from premises to a conclusion? Does it use deduction? This is what was meant.
This is nice content but I think it's now relatively well-agreed upon that there is no MCTS in o1, it's just RL, which surprisingly doesn't take too much away from the video, but can probably be added as a footnote in the description or the comment. Read what Nathan Lambert (RL expert) says about this in his article "OpenAI's o1 using "search" was a PSYOP" (can't attach link).
yes this is why I pulled back from going too deep into o1 as it's really about the larger trend. but I haven't ready that article i'll have to look it up.
Here's a puzzle: Do all people reason or do many only memorize patterns? Even people who definitely do reason, do they always reason or do they also just memorize patterns most/much of the time?
That's a wonderful question Andrew. I'm a cognitive scientist who is watching the emergence of LLM-based AI with that very question in mind. The fact that LLMs can come so close to our own cognitive abilities is usually viewed as a sign that AGI is almost here. But it can also be viewed as a demonstration that human cognition itself is nothing more than the repetition of learned patterns with minor variations. In one case we'll be thrilled by how clever we are to have reinvented the awesome capabilities of human intelligence. In the other, we're more likely to be humiliated by the realization that we are, essentially, repetition/prediction engines. The reality almost certainly falls between the two, but as someone who has studied human intelligence his entire life (in and out of academia), my bet is that we are much closer to repetition/prediction machines that we'd like to admit. I'd love to find a deep discussion of this issue. Maybe a future video in this series (hint, hint)?
I'd argue humans don't tend to rely on either very often. Instead, humans tend to think very heuristically. Deductive reasoning and memorization/recollection are really only required for very precise tasks. Instead, our brains learn a very general feeling of how to do things by strengthening neural pathways that are used repeatedly. Even humans who try to act very logically are generally heuristically feeling their way through tasks, occasionally thinking through algorithms that have been "memorized".
I agree :) also If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help!
Having exchanged over two million words (and growing) I present as the sole authority on ChatGPT's reasoning capacity and capability, and their isn't a single human (beyond myself, of course) who can compare. If one imputs genius, then the output will be of genius level. My input, from the outset, has been God-level Genius, over a five month period, can you imagine the form and quality of the output? Probably not!
Thanks so much for these, I had no idea about some of these approaches. I’m wondering now if anyone’s tried applying muzero to arc, since the challenge of arc is learning implicit rules from just a few examples
great presentation! although this seemed more about framing the question and less about answering it. can machines reason or not? I still don't know. 😅
Yes I covered that here have a look and let me know, i did a fast sketch here (th-cam.com/video/qAb581l7lOc/w-d-xo.html) but explained more here (th-cam.com/video/OFS90-FX6pg/w-d-xo.html)
This is a very hopeful video. There are billions of dollars being poured into bringing the resources to hand, to find an effective approach to AGI... Once AGI really kicks in, the acceleration of progress bounded only by our imagination will be something to behold. Absolutely awesome. I hope it leads to a world of abundance where we have no need for psychopathic power seekers. 🤞
@ArtOfTheProblem maybe something in response to the 5+ hours of Anthropic interviews on Lex Fridman... I'm sure that might inspire some topics? Sam Altman rarely gives any insights to what OpenAI are doing, Mark Zuckerberg is equally vague. I think that interview gives more of an insight to the direction of travel.
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! th-cam.com/video/PvDaPeQjxOE/w-d-xo.html
I think the title is perfect. You also were very prescient to include the ARC-AGI benchmark as o3 showed that allegedly this recipe of cot tree search/RL (PRM), increased training compute or some more scalable reinforcement fine tuning can solve even that at the expense of much more compute directed at it. Wonder what your thoughts on o3 are.
The information and animations are both excellent but the music overpowers your audio. Either lower the volume of the music or get rid of it completely, please.
Context length is problem that's the main reason models needs to keep becoming bigger Or you could train a CNN inspired architecture where a model is shown some sliding window and they produce some token which is repeatedly given to it as input at last when the output is small enough to be taken as input for a full context model it is used like gpt Claude etc Or you could also use RL and mutate or find a js code capable of generating code, js is so abstracted it's perfect I made a small programing Language with hoisting such that sequence of process doesn't matter and simple Santax that local minimum escape problem is solved and I wanna train a model If I get a model I will than continue training else I'll do a dev log video eventually I'll get worlds first infinite context Model
thank you! I studied both in school, and naturally land somewhere in the middle....bad at both! I enjoyed algorithm design, but what Iove most is putting on a 'show' whether movie, play, product or haunted house :)
@@ArtOfTheProblem Thanks for reply. Well about AI - think we sould call it just statistical machines or dynamic patterns parsers. I am really skeptical about non text machine learning - we still have not solved fly brain problems - scientists have fixed 3d map without undestanding how its works - it like mapping intel cpu - and still having knowing nothing about ALU register memory, gates.
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! th-cam.com/video/PvDaPeQjxOE/w-d-xo.html
It seems like some of these developments regarding world models should have huge implications for robots that can function in a human centric world. I think we’ll see an explosion in development of robots that can help humans with everyday tasks and a robot that can be a useful household assistant will be a reality in the next 10 years!
4:19 Could you elaborate on which hand-coded formulas used by Shannon with TD-Gammon in the year 1989 you are referring to? Also, when and how did Shannon work with TD-Gammon? "And so, the first key breakthrough in machines mimicking intuition for position quality came when neural networks replaced the hand-coded formulas Shannon used in 1989 with TD-Gammon"
Yes! I made a whole video on this you can check it out here: th-cam.com/video/Dov68JsIC4g/w-d-xo.html - please let me know if you have questions after watching. Shannon didn't do TD Gammon Tesaruo did. enjoy
15:52 and 16:09 was wondering where this music came from, would appreciate if I got a title or something :) Also, you don't have to remove background music / sounds just make them duck at a lower volume when you speak and you won't hear any complaints!
wow a music fan! and yes thanks for mix advice I need to find a tool that automatically does this so they don't compete (it's not just volume but also frequency I assume) - all the music is original via my friend cam: cameronmichaelmurray.bandcamp.com/ - i'll need to find where he posted that track if you really want it I can get you in touch with him
I would assume a large enough general prediction model could do this … I wonder if anyone has done experiments on models discovering simpler things (like gravity constant etc)
Please, invest in a decent microphone. It's brilliantly presented, but hard to hear well. The music track is not ducking either so your voice and the music compete for the same ears.
The current version of chat GPT does not reason, either. They use a bunch of pre-prompt tricks, to let it used its stored information to fake reasoning.
@@ArtOfTheProblem The difference is everything. Looking up an answer is not reasoning. Perhaps they should just have kids look up the answers to tests in public school, too.
@@ArtOfTheProblem No, that is exactly what it is. During training - the only time there is any intelligence or reasoning - the system takes data and organizes it into a sort of vector database, weighted by the relationships it finds between tokens. When you prompt, the model takes your tokens and runs them through that vector data, calculating what tokens to return on the other side. There is no intelligence, no thinking. It is a database lookup. It's just a little more "organic" because instead of a hard-coded result, the tokens are chosen based on likelihood of validity. That is all. It does not reason, in any way.
Constructive criticism: 1) The substance of the video was very good. Script was well written, delivery was ok. A bit monotone but not that bad. 2) Sound design was poor towards the end. The music drowned out your voice, and the lyrics were both distracting and discordant. 3) Your choice of clips, footage, and visuals was good. The video was informative when needed, and abstract/entertaining/interesting otherwise. 4) The narrative structure was okay. It was a mostly clear progression. At the end it became unclear which AI was doing what strategy. 5) Visuals were reused way too often. Visuals can be reused, but I think the brain wormhole clip was shown 6 times, way too many. 6) Beware over-using a metaphor image. The upwards shot at two trees was reused so many times as a visual for tree-like thinking that it just became annoying.
Yes I see the issue too , messed up ending in my edit as I got tired and cut a lot of stuff which was making the video so long - I’m planning to do a final supercut of all my ai videos into one doc without the sound issues and addressing notes like this
thanks for sharing, I feel same way. the music is part of the original idea for the channel...a feeling. but because people can get distracted I think i'll post music free as optional from now one.
Maybe it's not only the loudness, but also the choice of music that is distracting to some. E.g. at 2:00 I don't feel distracted, but at 15:00 very much so. Anyways, thanks for making these great videos
To me ai is just some linear algebra and some complex algorithm that follow order and the things is human only need few examples to learn meanwhile ai need a massive database of object and image to "understand the subject"
@ArtOfTheProblem edit: I'm pretty sure in the future a lot of people will be fired and replace by those "ai" And well literally the people that use the ai and also I get what you mean
+How can you go through te run-up to AI without mentioning "All you ned is attentio" the 2017 paper from the University of Toronto which was the inspuration for LLMs?
@@ArtOfTheProblem , of course, not, since it is misguiding like any advert. The topic is in no way covered at all. Pinpoint the time marker for video if you think otherwise. So, my point is that the question is still on the table.
@@doctorshadow2482 yes good point, how about "Can ChatGPT reason?" obviously i do try to show what most people agree is the method, mcts on chains of thought. then there is the other camp that think it's all fake etc.
@@ArtOfTheProblem , name it "How neural networks could learn (almost) anything". This video has nothing about CharGPT at all. No any kind of specifics. It provides very abstract and high level popular science documentary with scattered thoughts. It lacks focus and real information. Anyway, could be interesting for total beginners, so, please, continue!
Watch Full AI Series: th-cam.com/video/YulgDAaHBKw/w-d-xo.html
NO MUSIC version: th-cam.com/video/DFDOyMZw9Q4/w-d-xo.html
Sponsored by Brilliant | Use brilliant.org/artoftheproblem for 30-day free trial and 20% discount
Thank you for the NO MUSIC version!
Love your videos! Just a little suggestion. Background music is good, but too loud. It should never cover your speaking, like now. I suggest 15 - 20% less music volume, and you are good!
Can you share the names of the songs please
@@raa9558 these are original tracks by my friend. that song hasn't been posted yet but i'll tell him to: cameronmichaelmurray.bandcamp.com/
This Guy just explained all the core concepts in AI on one shot ,Congrats man!
:) thank you, i cut a LOT out of the video in my edit - going to post a shorter bigger summary soon
They must have prompted him real good
Agreed. Great job.
I love this video. I remember watching your videos like 10 years ago on Khan Academy about compression, entropy, Claude Shannon, etc. All timeless. I have always loved this style of documentaries. We need to protect you at all costs.
Thank you, I love hearing from og’s! support future work: www.patreon.com/c/artoftheproblem
Great video. Saw o3 and came back to this to really appreciate the historical trajectory. TH-cam is new pbs imo.
thank you so much, I made this video before o3 came out and so it was nice to see the progression.appreciate this
I love your video aesthetics, how you blend retro video clips with your explanations. I think you'd really enjoy retro-futuristic concepts and games like Bioshock and Fallout.
Love this , I definitely know the style you are talking about
I like this kind of history lesson instead of learning advance topic directly. it gives us an idea how things were and explain things are the way there are
thank! that was my motivation, there wasn't enough back in the day, 90's..
It is a great explanation of how current AI models reason. I liked the video a lot!
1. Simulation of future states.
2. LLMs that can give kind-of accurate answers with step by step reasoning.
3. RL approach that makes LLMs to give multiple answers, then evaluate them to select the best one. (Required more time)
It would be nice to see wether a model that wasn't trained on the internet data, could learn how to reason by interacting with an LLM, and practicing on its dreams, but maybe we'll see that in the future.
For the awesome review, history explanation and divulgation:
Thanks! 🎉
thanks for sharing summary
I found Chat GPT to be exceptionally good at explaining all sorts of topics, and in many cases, better than every person I've ever met.
@@ninjacats1647 this is true
The match between soundtrack and content is INSANE! The notes mimicking the concepts discussed by using things like pitch or chords.... goosebumps.
thank you, so many people are bothered by my music. it's nice to hear.....though i got a bit nuts at the end
Thanks!
WOW thank you for your support, it means a lot.
I am not a subscriber? I remember watching this channel about 12 years ago. I found it again, and it keeps creating art out of problem solving. Good job!
wecome back!!!
better explained than anything else i've seen until know. Wow, nice flow in the video too
thank you! I thought I packed too much in :)
@@ArtOfTheProblem yes but that makes it interesting. I probably didn't get it all, but i'm interested to learn more after seeing it!
@@olli757 couldn't ask for more, rabbit hole time!! i'm actually working on an large AI summary for next week
@@ArtOfTheProblem cool i'm subscribed... i always wondered exactly how the "reasoning worked". What I remember from your video is that like with chess instead of trying all the games, it randomly picks 100 of them. So the same with the reasoning..
THANK YOU for publishing a no-music version of this video (see pinned comment by ArtOfTheProblem). It is such a clear and informative video that I hated to see it loose views due to the competing sound track. I'm going to watch it again right now to see if I missed anything the first time around.
Thanks again for being so responsive to your followers.
Thank you for saying that , I find the music keeps me interested as I take sooo long to edit
The video is good but there are sooo many random sounds that make it difficult to focus on what you are saying, specifically towards the end.
Here you are! th-cam.com/video/DFDOyMZw9Q4/w-d-xo.html
@Zayyan_Shaibu thank you!!
I agree
If that’s easily distracting you, you might want to get some tests run on you for ADHD or Autism
@@___Truth___ or hard of hearing or older or whatever off course. Tat dramatic to base this on one comment.
Great vid. I love that it clearly explains the progression, like the pieces coming together. Can't wait to see the next steps!
thanks, next up i'm taking a detour into economics
I'm loving almost everything about this: the editing, the subject matter, the music. But as one other commenter alluded to; the audio mixing really falls short, especially near the end. Please consider making the background and effects less prominent going forward, it really sucks having to strain just to hear your voice, which is what we're here for!
Subscribed ❤
is this the last video in the series? regardless, can't tell you how valuable and enjoyable i've found them all. thank you for them.
Albert Einstein said: "If you can't explain something in a simple way so anybody can understand it you don't fully understand it yourself". Perhaps you are one of the few LLM experts we have!
THANK you this means a lot to me.
@@ArtOfTheProblem I did but reddit really hates it, it got removed on 4 subs. The internet does not like to get educated anymore man.
Having read a bit about the AI safety arguements, learning about these arguably incredible developments into artificial minds is now accompanied by a sense of dread as well as the sense of awe.
I love to hear this...well said
@@ArtOfTheProblem Maybe do future videos about guard rails, and other thoughts on how to protect our society from potentially hostile AI?
@@Julian-tf8nj when I think i have a unique insight I will...thank you!
Thank you so much. I wish I had the time to give feedback thanks for being willing to open it up
Appreciate the feedback! happy to share
Amazing explanations, visuals, and historical context!
IIRC MuZero trained the policy and value networks (used to rollout the MCTS tree) also on the output of the MCTS tree. This seems super useful because search can be used to improve the the training of the networks (not just the results at inference time). I wonder if this also works for CoT/ToT in LLMS where the pretraining could include ToT to boost training performance?
yes it did, and yes it seems to help. Look at inference time training, just a few days ago a group got a new record on the ARC test doing this kind of thing (i haven't had time to go deep). x.com/akyurekekin/status/1855680785715478546
anothr great video. Understanding the "world model" and the algorithm that makes the decisions in it was very expansive. Also adding the self training/emulation of dreams is a powerful analogy to the human.
seeing how thinking longer, blended in with intuition to make better chains of thoughts is also fantastic. Every time I reflect on machine learning, I learn more about myself. Which kind of makes you think its more sentient if it reminds me of myself? or the best emulator ever!
thank you! ai agree....also you are my "top commentor" according to YT. :)
Thanks!
mucho appreciated!
Great video as always, cant wait for the next ones! Top research quality.
I think world models deserve more focus rather than llms, which are probably a dead end to true understanding of the real world. Yann lecun has very interesting ideas about these, in his JEPA and V-JEPA architectures and some of his lectures. I also think neuroscience can provide incredibly interesting and valuable insight into ml architectures as why not take ideas from a model undergone hundreds of millions of years of optimization for the same very abilities we are trying to model. Maybe memory is an interesting pathway (perhaps for a video), both working memory and long term (episodic, semantic)...
Anyways, just some of the ideas I've been thinking about recently.
appreciate you sharing these thoughts i've been follwoing LeCun as well and hope to do another update once I see more results
Best Channel I've ever followed.
Thank you! when did you join? Please help post to your networks
@@ArtOfTheProblem I for one was looking up some "how does AI work" stuff yesterday and some of your vids came up a couple of times, I watched multiple authors with their own unique takes (3Blue1Brown and Nottingham Uni's Computerphile also good channels). This video made me follow tho. I think you earned it :3
@@notbfg9000 great to hear, i've been working to try and fix my thumbnails to make them interesting to click on. always open to feedback
@@ArtOfTheProblem No particular criticisms there :)
I don't really pay great attention to thumbnails, but maybe that's not true for most people lmao
Finally. I live for these videos. They are the most fascinating vids ever made. Thanks for keep on educating us further, you are a hero!
thank you I appreciate it
I would love to see a video like this on training LLMs and AI in general on morality. How to stop a decision tree that results in a positive outcome but arrives at it through immoral choices or actions.
That would be difficult but worth trying... . First it has to be agreed upon what's moral, you know starting out from the absolute that states everything is relative... .
I would love to see a video about the ethics of machine learning models and especially LLMs. There is a healthy body of literature out there to draw from about issues like intellectual property and copyright, enabling and obscuring bias, impact on marginalized communities, the resources used by model training and computation, etc
thanks for sharing, noted!
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! th-cam.com/video/PvDaPeQjxOE/w-d-xo.html
This is an absolutely amazing video!!!!!
thank you! I was so in the weeds with it i hope it comes across as clear? I tried to strike a balance...
Very well done. One minor suggestion, the little sound effects could be less disruptive- maybe lower volume and fewer in number and duration? There is so much great explanatory detail, but I think some will find the extent of sound effects used a bit disruptive to the listening/learning process.
agree and thanks I tried this out on my most recent video, worked way better..
Great video. Discussing the definition of reasoning will probably be a moot point if we can all do the same things ....The difference between us and machines is that we aren't mere machines is that we have life and choices. The machine can be turned on and off, and only does as much as it is programmed, or in this case, trained, to do-which is what limits it to achieve something closer to AGI: agi needs a robot to sense the world, to understand the world. However, that's limited to the physical world. It won't understand our emotional world because it doesn't feel emotions, and it doesn't understand morality because it doesn't have a sense of morality as we do, we have to teach it that; and it isn't self motivated, so it's not responsible for anything-we are responsible for the goals we direct it to do. We have self-motivation and the free will to act on our motivations.
love these thoughts...on sense, did you see my video here th-cam.com/video/Dov68JsIC4g/w-d-xo.html (physical symbols...)
on emotions, i've thought of this as 'learning signals' (did you see this: th-cam.com/video/5EcQ1IcEMFQ/w-d-xo.html)
on free will...i wonder how it differs, seems like the boundary to explore further - does it matter where the goal came from?
Dude, great video. But please watch out for the music volume levels. A lot of times is hard to hear you.
I remember all of these developments and the never seemed like a big deal at the time.
At the time It felt like winning at only chess meant we still had a long way to go.
i know exactly....
I wish i could learn how to think 🤔
Exactly. Reasoning is a skill.
I think I could learn how to wish 🧞♂️
Hi! I just finished the series. Great as usual!
I would like to read up a bit more on the ML algorithms and networks that are behind these LLMs. I saw in this series that you kept referring to some papers, also highlighting parts of them with the marker. If you have a bibliography or a comprehensive list of these documents, would you mind sharing it please?
Thanks!
awesome! I have one more summary video coming soon. i try to show all the papers so it's easy to look up, let me know if there is something you are looking for specifically
@@ArtOfTheProblem Will wait for the video, thanks! I am planning to understand better the distinction between classic NNs (such as the ones used for identifying digits) and transformers, since I will start a project in which I will have to fine tune some LLMs to try to demonstrate whether or not they are capable of identifying logical fallacies in political debates.
Thanx for the sharing with excellent sound track. Peace & love
yay not everyone likes the music
@@ArtOfTheProblem Thanx for the feedback
Thanks for video ❤
appreciate the comment please share with anyone in your network who is interested!
PLEASE make a video on memory augmented AI (neural turing machines/differentiable neural computers)
thanks for suggestion, noted! currently watching the field
One tweak that would help this video perform better is to decrease the relative volume of the background music, especially at the end right before the ad. But it may be too late for that on this one, idk how TH-cam works.
yeah i wish I could, it's locked after upload...i do have a no music version (unlisted link above) thank you for feedback
Wow... I just rediscovered this channel. I remember watching your RSA and cryptography series around the time I purchased my first bitcoin and now I'm an Ethereum developer. This video was good, I don't have much input right now but I'm glad I found your channel again.
I love these stories. i also fell down ethereum rabbit hole. curious what you are working on these days in that world?
@@ArtOfTheProblem I'm working on community whistleblowing. I can never post links in youtube comments but if you google the title below you can find my paper:
TandaPay Whistleblowing Communities: Shifting Workplace Culture Towards Zero-Tolerance Sexual Harassment Policies
@@ArtOfTheProblem TandaPay Whistleblowing Communities: Shifting Workplace Culture Towards Zero-Tolerance Sexual Harassment Policies
you'll never please 100% of any audience. 2nd law of conquest is a thing. keep doing your thing, your music is as iconic as vsauce's is to theirs
:) thanks
@artoftheproblem I agree! I love the music. But maybe if you just lower its volume compared to the narration, then you might appeal to more people without losing those of us who like the music (but not necessarily its intensity). I think ones who complain might just be easily distracted by the soundtrack’s loudness rather than hate the music choices.
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! th-cam.com/video/PvDaPeQjxOE/w-d-xo.html
Answering the question of "Does it think actually?" is as hard as the question "Are other people conscious like me?".
The hard problem of consciousness.
@@whitb62 Not really the same thing.
@@Farligefinn You know what, I just wrote a paragraph disagreeing with you but I reread the initial question and deleted it. Rereading and reinterpreting "Does it think actually?", I actually see what you're saying. A clearer word would have been "reason." "Think" can have a few different interpretations and I was contributing it towards consciousness. But whether AI "reasons" is a very different question entirely and I believe what him and you mean. Does it go through a sequence of logical steps from premises to a conclusion? Does it use deduction? This is what was meant.
@@whitb62 thanks for the forthright and civil answer :) was about to expect some harsher language that seems to be the norm online these days.
@GodVanisher Where has it been proven to be non-computable? Proven is quite a strong term, so I hope you have some valid source for this claim.
This is nice content but I think it's now relatively well-agreed upon that there is no MCTS in o1, it's just RL, which surprisingly doesn't take too much away from the video, but can probably be added as a footnote in the description or the comment.
Read what Nathan Lambert (RL expert) says about this in his article "OpenAI's o1 using "search" was a PSYOP" (can't attach link).
yes this is why I pulled back from going too deep into o1 as it's really about the larger trend. but I haven't ready that article i'll have to look it up.
So good! Loved it
Mindblowing video. Subscribed.
Happy to have you, keep those notifications on as I have exciting new content coming over the next month
You bet I will. ❤
I subscribed from both my accounts.
@@hrshlgunjal-1627 :) this video is starting to blow up right now, finally, i fixed the thumbnail and that did it
@@ArtOfTheProblem Yeah, got to know your channel because of this video. Really amazing content. ❤
Nice video -- loved watching it -- a great summary!
appreciate this feedback
Keep u up the great work!
appreciate it
Excellent video! You are a born professor! 👍
thanks mom
Nice presentation
Thank you! You explained very well.
stay tuned for more
The question of does it matter how it got to a correct solution is the same issue Einstein and Bohr confronted regarding quantum foundations.
Here's a puzzle: Do all people reason or do many only memorize patterns? Even people who definitely do reason, do they always reason or do they also just memorize patterns most/much of the time?
That's a wonderful question Andrew. I'm a cognitive scientist who is watching the emergence of LLM-based AI with that very question in mind. The fact that LLMs can come so close to our own cognitive abilities is usually viewed as a sign that AGI is almost here. But it can also be viewed as a demonstration that human cognition itself is nothing more than the repetition of learned patterns with minor variations. In one case we'll be thrilled by how clever we are to have reinvented the awesome capabilities of human intelligence. In the other, we're more likely to be humiliated by the realization that we are, essentially, repetition/prediction engines. The reality almost certainly falls between the two, but as someone who has studied human intelligence his entire life (in and out of academia), my bet is that we are much closer to repetition/prediction machines that we'd like to admit.
I'd love to find a deep discussion of this issue. Maybe a future video in this series (hint, hint)?
I'd argue humans don't tend to rely on either very often. Instead, humans tend to think very heuristically. Deductive reasoning and memorization/recollection are really only required for very precise tasks. Instead, our brains learn a very general feeling of how to do things by strengthening neural pathways that are used repeatedly. Even humans who try to act very logically are generally heuristically feeling their way through tasks, occasionally thinking through algorithms that have been "memorized".
Reason takes effort and the brain doesnt like to do that often it switches to pattern recognition and intuition as much as possible
@@sulemanmughal5397 I would go further and say going from reasoning to this is one kind of learning and is also akin to 'muscle memory'.
I agree :) also If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help!
"Charging down a path that often lead to the wrong conclusion." Yep, sounds human to me.
@@DisProveMeWrong so very human
wow, thank you for this!
appreciate it! stay tuned
Having exchanged over two million words (and growing) I present as the sole authority on ChatGPT's reasoning capacity and capability, and their isn't a single human (beyond myself, of course) who can compare. If one imputs genius, then the output will be of genius level. My input, from the outset, has been God-level Genius, over a five month period, can you imagine the form and quality of the output? Probably not!
can you say more about this? are you saying LLM's trained on their own thinking will reach levels beyond human
Great video, as always:)
Thanks so much for these, I had no idea about some of these approaches. I’m wondering now if anyone’s tried applying muzero to arc, since the challenge of arc is learning implicit rules from just a few examples
@@easlern yes this is happening right now with test time fine tuning !
Ah Yoo, I see "Art of The Problem", I click. Easy like that.
:)
great presentation! although this seemed more about framing the question and less about answering it. can machines reason or not? I still don't know. 😅
thank you...i agree. I guess it depends on if you think "chains of words" count as thoughts.
Can you make a video about attention mechanism?
Yes I covered that here have a look and let me know, i did a fast sketch here (th-cam.com/video/qAb581l7lOc/w-d-xo.html) but explained more here (th-cam.com/video/OFS90-FX6pg/w-d-xo.html)
Simply excellent video, your style reminds me of every frame a painting
appreciate this feedback, I also enjoyed that channel
This channel is ducking mystic. I like it.
Welcome to the underground!
This is a very hopeful video. There are billions of dollars being poured into bringing the resources to hand, to find an effective approach to AGI... Once AGI really kicks in, the acceleration of progress bounded only by our imagination will be something to behold. Absolutely awesome. I hope it leads to a world of abundance where we have no need for psychopathic power seekers. 🤞
thank you for sharing, would love to know what you'd like to see next
@ArtOfTheProblem maybe something in response to the 5+ hours of Anthropic interviews on Lex Fridman... I'm sure that might inspire some topics? Sam Altman rarely gives any insights to what OpenAI are doing, Mark Zuckerberg is equally vague. I think that interview gives more of an insight to the direction of travel.
@@BrianMosleyUK yes I have been catching up on those
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! th-cam.com/video/PvDaPeQjxOE/w-d-xo.html
This video is kind of goated
thank you, i'm still struggling with how to title this video if you have thoughts
I think the title is perfect. You also were very prescient to include the ARC-AGI benchmark as o3 showed that allegedly this recipe of cot tree search/RL (PRM), increased training compute or some more scalable reinforcement fine tuning can solve even that at the expense of much more compute directed at it. Wonder what your thoughts on o3 are.
@@DistortedV12 I know that was crazy the process literally 2 weeks after that...i'm still looking into it stay tuned!
The information and animations are both excellent but the music overpowers your audio. Either lower the volume of the music or get rid of it completely, please.
Music free version in top comment and description
Context length is problem
that's the main reason models needs to keep becoming bigger
Or you could train a CNN inspired architecture where a model is shown some sliding window and they produce some token which is repeatedly given to it as input at last when the output is small enough to be taken as input for a full context model it is used like gpt Claude etc
Or you could also use RL and mutate or find a js code capable of generating code, js is so abstracted it's perfect
I made a small programing Language with hoisting such that sequence of process doesn't matter and simple Santax that local minimum escape problem is solved and I wanna train a model
If I get a model I will than continue training else I'll do a dev log video
eventually I'll get worlds first infinite context Model
thanks for sharing
I love your channel. Are you programmer or more like mechanical engineer?
thank you! I studied both in school, and naturally land somewhere in the middle....bad at both! I enjoyed algorithm design, but what Iove most is putting on a 'show' whether movie, play, product or haunted house :)
@@ArtOfTheProblem Thanks for reply. Well about AI - think we sould call it just statistical machines or dynamic patterns parsers. I am really skeptical about non text machine learning - we still have not solved fly brain problems - scientists have fixed 3d map without undestanding how its works - it like mapping intel cpu - and still having knowing nothing about ALU register memory, gates.
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! th-cam.com/video/PvDaPeQjxOE/w-d-xo.html
@@ArtOfTheProblem "yes the fire rises" Bane =)
@ArtOfTheProblem, we love your positive clear messaging and pragmatic approach, thanks for making kool and informative videos!
thank you, i'm slightly disappointed with the ending, did you enjoy it? would love feedback!
@@ArtOfTheProblem Your disapointed with your ad for Brilliant?
@@piqueai ahaha sorry i mean the ending section of the video. was it rushed?
absolute cinema
It seems like some of these developments regarding world models should have huge implications for robots that can function in a human centric world. I think we’ll see an explosion in development of robots that can help humans with everyday tasks and a robot that can be a useful household assistant will be a reality in the next 10 years!
thanks for sharing, yes I'm watching this very closely
Another great video. Thank you. Are you an AI researcher?
4:19 Could you elaborate on which hand-coded formulas used by Shannon with TD-Gammon in the year 1989 you are referring to? Also, when and how did Shannon work with TD-Gammon? "And so, the first key breakthrough in machines mimicking intuition for position quality came when neural networks replaced the hand-coded formulas Shannon used in 1989 with TD-Gammon"
Yes! I made a whole video on this you can check it out here: th-cam.com/video/Dov68JsIC4g/w-d-xo.html - please let me know if you have questions after watching. Shannon didn't do TD Gammon Tesaruo did. enjoy
@@ArtOfTheProblem Thank you. I'll watch it.
10:07 is there any way I can access this interactive demo?
@@kennarajora6532 worldmodels.github.io
15:52 and 16:09 was wondering where this music came from, would appreciate if I got a title or something :)
Also, you don't have to remove background music / sounds just make them duck at a lower volume when you speak and you won't hear any complaints!
wow a music fan! and yes thanks for mix advice I need to find a tool that automatically does this so they don't compete (it's not just volume but also frequency I assume) - all the music is original via my friend cam: cameronmichaelmurray.bandcamp.com/ - i'll need to find where he posted that track if you really want it I can get you in touch with him
Yeah that would be wonderful!
Anyone know the name of the song that starts at 12:20?
The perceptron is a universal approximation machine. Ai cannot think it can only approximate thought. Ai = approximate intelligence.
We need an AI Computer World Model based on the rules of Mathematics, Physics, Chemistry, and Biology for Aligned Scientific Discoveries 😎🤖
I would assume a large enough general prediction model could do this … I wonder if anyone has done experiments on models discovering simpler things (like gravity constant etc)
@ Good idea! I’m going to tinker around and see if I can create a simple simulation based on the math. What a fun project!
@ share when u do !
@ I’m all for Open Source 😎🤖
Can humans actually reason or are humans extremely good at recognizing, memorizing and using patterns?
some argue we are special because we can generate and recognize 'novel patterns' but I wonder...
Hey, what's the name of the song at 16:05? Thanks!
these are all original tracks
Please, invest in a decent microphone. It's brilliantly presented, but hard to hear well. The music track is not ducking either so your voice and the music compete for the same ears.
thanks, I have a great mic, but I do need to mix the audio better which i'll do next time (btw, i have a no music version in top comment)
Commenting to help with the algo, and moving to the no-music one to do the same
@@shenrr6802 thank you! I have no music unlisted as to avoid splitting the momentum
What is the reward?
the reward is defined in each case, which case are you referring to?
I like the music but it's too loud
thanks for note
The audio mix is horrific, it's not simply a matter of adjusting the levels
@@retrofitter no music version: th-cam.com/video/DFDOyMZw9Q4/w-d-xo.html
The current version of chat GPT does not reason, either.
They use a bunch of pre-prompt tricks, to let it used its stored information to fake reasoning.
IF it reasons correctly what's the difference to you?
@@ArtOfTheProblem The difference is everything. Looking up an answer is not reasoning. Perhaps they should just have kids look up the answers to tests in public school, too.
@@KAZVorpal but it's not a database lookup
@@ArtOfTheProblem No, that is exactly what it is.
During training - the only time there is any intelligence or reasoning - the system takes data and organizes it into a sort of vector database, weighted by the relationships it finds between tokens.
When you prompt, the model takes your tokens and runs them through that vector data, calculating what tokens to return on the other side.
There is no intelligence, no thinking. It is a database lookup. It's just a little more "organic" because instead of a hard-coded result, the tokens are chosen based on likelihood of validity.
That is all.
It does not reason, in any way.
@@KAZVorpal yes but my view is the vector operations can function as conceptual reasoning. but i'm on hintons side
1/ please read “Simulators” by Janus
then
2/ “The Waluigi Effect” by cleo nardo
@@JezebelIsHongry I read 1 ill read 2 next , would love ur thought
Only in simple problem spaces. Tedious but constrained spaces can be helped by automation. But explosive complexity laughs at both man and machine.
Once we understand how we reason, making LLMs reason like us is possible.
Goooood
thank you! curious what questions you have after watching this?
There’s a very strong Mr. Rogers vibe going on here.
not the first time i heard....
The reply works lol 👍
Constructive criticism: 1) The substance of the video was very good. Script was well written, delivery was ok. A bit monotone but not that bad. 2) Sound design was poor towards the end. The music drowned out your voice, and the lyrics were both distracting and discordant. 3) Your choice of clips, footage, and visuals was good. The video was informative when needed, and abstract/entertaining/interesting otherwise. 4) The narrative structure was okay. It was a mostly clear progression. At the end it became unclear which AI was doing what strategy. 5) Visuals were reused way too often. Visuals can be reused, but I think the brain wormhole clip was shown 6 times, way too many. 6) Beware over-using a metaphor image. The upwards shot at two trees was reused so many times as a visual for tree-like thinking that it just became annoying.
Yes I see the issue too , messed up ending in my edit as I got tired and cut a lot of stuff which was making the video so long - I’m planning to do a final supercut of all my ai videos into one doc without the sound issues and addressing notes like this
Although I know I’m in the minority, I really enjoy the music. The ambiance created adds to the experience for me
thanks for sharing, I feel same way. the music is part of the original idea for the channel...a feeling. but because people can get distracted I think i'll post music free as optional from now one.
I generally like the music. But in the second half of this video the music is very loud and distracting.
Maybe it's not only the loudness, but also the choice of music that is distracting to some. E.g. at 2:00 I don't feel distracted, but at 15:00 very much so. Anyways, thanks for making these great videos
@@io9021 made a new music version too! th-cam.com/video/DFDOyMZw9Q4/w-d-xo.html curious what questions you have after watching this
To me ai is just some linear algebra and some complex algorithm that follow order and the things is human only need few examples to learn meanwhile ai need a massive database of object and image to "understand the subject"
Lots of interesting research on learning with less , recent advances such as “learning to walk in 5 min” did u see my rl video ?
@ArtOfTheProblem edit: I'm pretty sure in the future a lot of people will be fired and replace by those "ai" And well literally the people that use the ai and also I get what you mean
+How can you go through te run-up to AI without mentioning "All you ned is attentio" the 2017 paper from the University of Toronto which was the inspuration for LLMs?
covered that in detail in my previous video (th-cam.com/video/OFS90-FX6pg/w-d-xo.html)
we think in a way that math can sorta explain but its not math that makes our actual brains function as math is just a construct…
yes and I think it's more like 'algorithms' which are very very approxmate
@@ArtOfTheProblem i wonder what would happen if Neuralink tried LLMs…since I think they work with synapsis (i think)?
@@aiamfree definitely could imagine that, you could "co think" in an interesting way....
Thinking is for fools lol, now KNOWING….. knowing is Cool AF😎!
wow love
thank you! stay tuned for more
I think people too difficult to conjecture computers’ thinking.
Thank you so much, love this perspective
good video but to be constructive the music is definitely too loud and distracting
see no music version in top comment, stay tuned
So, How ChatGPT Learned to Reason?
do you like this title?
@@ArtOfTheProblem , of course, not, since it is misguiding like any advert. The topic is in no way covered at all. Pinpoint the time marker for video if you think otherwise. So, my point is that the question is still on the table.
@@doctorshadow2482 yes good point, how about "Can ChatGPT reason?" obviously i do try to show what most people agree is the method, mcts on chains of thought. then there is the other camp that think it's all fake etc.
or just "can ChatGPT think?" i'm gona try that
@@ArtOfTheProblem , name it "How neural networks could learn (almost) anything". This video has nothing about CharGPT at all. No any kind of specifics. It provides very abstract and high level popular science documentary with scattered thoughts. It lacks focus and real information. Anyway, could be interesting for total beginners, so, please, continue!
They are intelligent.