Alpha Everywhere: AlphaGeometry, AlphaCodium and the Future of LLMs

แชร์
ฝัง
  • เผยแพร่เมื่อ 17 ม.ค. 2024
  • Is AlphaGeometry a key step toward AGI? Even Deepmind's leaders can't seem to make their minds up. In this video, I'll give you the rundown of what AlphaGeometry is, what it means and what it doesn't meann. Plus I'll cover AlphaCodium, dropped open-source tonight seemingly out of nowhere, and causing a big stir for what it might mean for coders the world over. And I'll touch on what I foresee is the future of large languages models and their alliance with search.
    AI Insiders: / aiexplained
    AlphaGeometry: www.nature.com/articles/s4158...
    Deepmind Blog Post: deepmind.google/discover/blog...
    Shane Legg Tweet: / 1747670093348176140
    AlphaCodium Paper: arxiv.org/pdf/2401.08500.pdf
    Alpha Codium Blog: www.codium.ai/blog/alphacodiu...
    Alpha Codium … Code: github.com/Codium-ai/AlphaCodium
    AlphaCodium Tweets: / 1748043513156272416
    / 1747971746047627682
    Twitter Math Kardashian: / 1746168116546093291
    Hassabis New Tweet: demishassabis/sta...
    AlphaGeometry NYT: www.nytimes.com/2024/01/17/sc...
    AIMO Prize: aimoprize.com/
    Metaculus IMO: www.metaculus.com/questions/6...
    Paul Christiano Lesswrong: www.lesswrong.com/posts/sWLLd...
    Professor Rao, On the Planning Abilities of LLMS: arxiv.org/pdf/2305.15771.pdf
    Eureka: HUMAN-LEVEL REWARD DESIGN VIA CODING LARGE LANGUAGE MODELS arxiv.org/pdf/2310.12931.pdf
    Google Offers Salary: www.theinformation.com/articl...
    Mathematics Will Fall First: / 1731096582932578653
    Samsung Galaxy S24 Gemini Ultra and Nano: blog.google/products/android/...
    Lead Author Trieu Video: • AlphaGeometry
    V100 to X100: pbs.twimg.com/media/FbkEJX1WY...
    substackcdn.com/image/fetch/w...
    AI Explained Eureka: • State of AI 2023: High...
    Check out the amazing Donato Capitella: / @donatocapitella Non-Hype, Free Newsletter: signaltonoise.beehiiv.com/
  • วิทยาศาสตร์และเทคโนโลยี

ความคิดเห็น • 290

  • @DaveShap
    @DaveShap 4 หลายเดือนก่อน +428

    Okay I really want Robin Williams to teach me math.
    Update: but also getting AI to master math is absolutely critically accelerating us to AGI.

    • @DaveShap
      @DaveShap 4 หลายเดือนก่อน +86

      ​@@person737 that is a comically misguided statement by him.

    • @dustinbreithaupt9331
      @dustinbreithaupt9331 4 หลายเดือนก่อน +8

      Who invited this guy?😁

    • @DaveShap
      @DaveShap 4 หลายเดือนก่อน +49

      Inventors rarely understand the full impact of their intentions. Especially in the long run.

    • @ryzikx
      @ryzikx 4 หลายเดือนก่อน

      david shapiro is the most correct predictions, i still believe biological immortality by 2030

    • @jumpstar9000
      @jumpstar9000 4 หลายเดือนก่อน +21

      ​@person737 From everything I have seen Sama tends to misdirect. He's like a magician and magicians assistant all rolled into one. Wasn't it at Dev Days He said the stuff they were working on would make everything so far seem quaint. I think he is trying hard to keep a lid on things and preserve the moat. Well imho 🤷‍♂️

  • @GrindThisGame
    @GrindThisGame 4 หลายเดือนก่อน +155

    As different things are bolted together it reminds me of various regions of the brain that are specialized and work together. Thanks again for a great video.

    • @waitean09
      @waitean09 4 หลายเดือนก่อน

      I always thought that'd be how we achieve AGI.

    • @Enderlad
      @Enderlad 4 หลายเดือนก่อน +1

      love the way you put it and I have thought the same.

    • @dewyocelot
      @dewyocelot 4 หลายเดือนก่อน +5

      This is what I think about when people say AI can’t do X. The emergent properties that can (and have) come from these parts combining and working together is nuts.

  • @pydron
    @pydron 4 หลายเดือนก่อน +29

    It's always a good day when you see that black text on a white background
    Great video!

  • @wealthycow5625
    @wealthycow5625 4 หลายเดือนก่อน +116

    Best AI channel on TH-cam, hands down. Been following you since the start!

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน +20

      Oh wow thanks cow!

    • @toshmosh
      @toshmosh 4 หลายเดือนก่อน +1

      it's because you stick to the research instead of hype-y things and don't dumb things down! @@aiexplained-official

    • @TSSPDarkStar
      @TSSPDarkStar 4 หลายเดือนก่อน +5

      yeah for people who actually want to dig into the details at a decently approachable level 100%

    • @iamcoreymcclain
      @iamcoreymcclain 4 หลายเดือนก่อน +1

      10:06 it’s becoming more like a conversation between LLMs and their environments.
      Something I have recently started doing about a month ago, is taking a narrative from my life or someone else’s when I am using ChatGPT and placing ChatGPT inside of that narrative for the totality of that conversation. I’m not talking about telling it to act as a copywriter or systems engineer, but actually giving it a full story with emotional context of the human characters it is going to be helping.
      I haven’t tested this in a scientific way, but it seems that the more narrative I provide the more productive the conversation is.
      I’m not sure if there is even any connection but what you said made me think about how I’ve been using it and how happy I’ve been with the results personally

  • @QuickM8tey
    @QuickM8tey 4 หลายเดือนก่อน +33

    I really appreciate how you always leave citations in the description. This is easily one of the most transparent and well researched channels I watch. Really excited to see how the rest of 2024 plays out with AI developments.

  • @londonspade5896
    @londonspade5896 4 หลายเดือนก่อน +89

    This channel is like setting up a lawn chair on a highway, sitting comfortably and watching a massive truck coming closer and closer 🍿

    • @vladgheorghe4413
      @vladgheorghe4413 4 หลายเดือนก่อน

      Nailed it.

    • @devinfleenor3188
      @devinfleenor3188 4 หลายเดือนก่อน

      It's interesting to think that if an asteroid was gonna hit earth people would probably just vibe.

  • @cacogenicist
    @cacogenicist 4 หลายเดือนก่อน +65

    I envision systems with a bunch of specialist-ish modules, bolted together the right way, with some sort of integrational module routing inputs to the appropriate domain specialist modules, with an attentional/executive component steering the whole thing.
    And eventually it would be nice to approximate the plasticity of brains with ongoing updating of the nets.

    • @ZoOnTheYT
      @ZoOnTheYT 4 หลายเดือนก่อน +8

      That's as good a definition of consciousness as any other accepted ones that I've heard.

    • @jebprime
      @jebprime 4 หลายเดือนก่อน +1

      I think specialization modules would provide a sort of statistical bias for the models that allow them to learn faster or more accurately than any single module trained to perform a task.

    • @MrHuman002
      @MrHuman002 4 หลายเดือนก่อน +4

      That's more or less exactly what MoE (Mixture of Experts) is, which is how Mixtral 8x7B works, and how many suspect OpenAI works.

    • @Kazekoge101
      @Kazekoge101 4 หลายเดือนก่อน

      I think he means a more complicated and higher level structure that has some self-similarity to MoE I guess@@MrHuman002

    • @antman7673
      @antman7673 4 หลายเดือนก่อน

      I don’t think that will be the path. LLMs are already, everything experts.
      Just the right prompt and you can access an expert.

  • @SnapDragon128
    @SnapDragon128 4 หลายเดือนก่อน +15

    Speaking as an IMO Silver medallist, I salute my new robot overlord. But seriously, I've long wanted humans to be able to convert their body of math knowledge into computer-verifiable symbolic form, so at the very least I'm excited for the prospect of LLMs being able to do that for us.
    Also, the real advantage of solving math problems symbolically is that there's no hallucination issue - the system knows whether it's gotten things right or wrong. As the DeepMind folks pointed out, the system might not come up with an _elegant_ proof, but it will always be a _correct_ proof. That's awesome!

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน +2

      Great to have you here Snap

    • @ShawnFumo
      @ShawnFumo 4 หลายเดือนก่อน +1

      And I’m sure eventually they can be trained to prefer elegance. Same way image generation is fine-tuned for human aesthetics.

  • @MadeOfParticles
    @MadeOfParticles 4 หลายเดือนก่อน +6

    AlphaGeometry and Eureka demonstrate that LLMs are already conducting their own form of semi-research. Many people don't recognize this, but LLMs are essentially AGI, only requiring enhanced reasoning abilities to solidify this, akin to a child maturing into an adult. For those who perceive LLMs as merely advanced word-predicting semantic algorithms, I suggest trying to learn a new language and constructing novel sentences. You'll realize the effort involved in predicting the next word in order to form meaningful sentences. This challenge is often overlooked because, as children learning a language, we struggled to form coherent sentences. We trained repeatedly until we could recall the patterns in language that made sense. Similarly, LLMs can reason; they just need to deepen this ability. Once that occurs, LLMs will likely solve mathematical problems on the first attempt and be capable of generalizing mathematics. At the end of the day, math is a language for understanding natural order.

  • @themonsterintheattic
    @themonsterintheattic 4 หลายเดือนก่อน +12

    AI Explained and Dave
    Shapiro are by far the best channels dealing with AI right now

    • @esuus
      @esuus 4 หลายเดือนก่อน

      Yeah but Shapiro is not a close second.

  • @kickmonlee3390
    @kickmonlee3390 4 หลายเดือนก่อน +7

    Kim telling taylor that the "dx" was way too thick is some AI comedy gold! Nah Kim its not thick enough and we all know you could make it thicker.

  • @jvlbme
    @jvlbme 4 หลายเดือนก่อน +10

    Don't really know what to think of all the "AGI is near", "well, not THAT near", or "AI will take all human jobs", "I said that? I just meant _transform_ jobs" statements these days - and these conflicting comments are from the SAME people... Backtracking is the way of 2024 apparently. 🤷

    • @jumpstar9000
      @jumpstar9000 4 หลายเดือนก่อน +3

      Even if it were here today, which depending on your perspective some might say is the case, it takes time for the world to change. It isn't like a light switch and everything changes. We can probably look back at the first few Industrial Revolutions for guidance. There is always a buffer period. Society is going to need to adapt, and a lot of infrastructure still needs putting in place. That means digging up roads and building data centers, education. All manner of things that take time. It is going to look quite soft for a while yet.

    • @involuntaryoccupant
      @involuntaryoccupant 4 หลายเดือนก่อน +1

      it's hard to predict what new tech will come out tomorrow. agi means above human intelligence, and we can't define what human-level intelligence is still. what i think is clear for now is that if you work with software (programmer, developer, even banking and accounting) you better start learning how to use ai as a force multiplier to not get laid off

  • @Dannnneh
    @Dannnneh 4 หลายเดือนก่อน +16

    Indeed, an AI that can work as a kind of motherboard to manage and interface between several specialized AIs would be interesting to see how it would perform.

  • @cuentadeyoutube5903
    @cuentadeyoutube5903 4 หลายเดือนก่อน +3

    Regarding detailed prompting, I’m starting to create a computer game completely assisted by LLM (gpt4, Claude, bard, dalle). And the first thing I found out was that not skipping steps gets me much far. I.e. if I go from “idea to code” I fail completely. Instead, I go from “brainstorm” to “concrete idea” to “roadmap” to “vertical slice tickets in first version” to “architecture discussion and definition” to “ actionable tickets” to “help me implement these tickets”. Always giving relevant context I got from the previous steps to achieve results on the next step. This is done manually for now, but I guess if I find this working I’ll wrap it up into a more formal product.

  • @jiucki
    @jiucki 4 หลายเดือนก่อน +5

    Alpha geometry has part of the code public in a repo! That's very interesting coming from Google 😮

  • @fynnjackson2298
    @fynnjackson2298 4 หลายเดือนก่อน +7

    AI is not going anywhere, things are heating up, crazy how much cash, energy and passion is now behind developing and sharing this, future gonna be absolutly mind-blowing in all the best ways.

  • @williamjmccartan8879
    @williamjmccartan8879 4 หลายเดือนก่อน +1

    Thank you Phillip, it seems like you're prompting of the LLM'S last spring in the way you did was on target not just with this news, but some of the things covered in the QStar video, have a great night, and thank you for sharing your time and work, peace

  • @alpha007org
    @alpha007org 4 หลายเดือนก่อน +6

    Thank you for providing up to date, quick and concise videos.

  • @Ecthelion3918
    @Ecthelion3918 4 หลายเดือนก่อน +2

    Just started reading about this today, very excited to hear your thoughts

  • @shawnryan3196
    @shawnryan3196 4 หลายเดือนก่อน +4

    I know this is harder than it sounds but I really believe what will make things happen is - Giving the model continuous updatIng to the world model while updating a "belief" system with RL.

  • @rioiart
    @rioiart 4 หลายเดือนก่อน +9

    Yay, a new AI explained video!

  • @NoahGibson1
    @NoahGibson1 4 หลายเดือนก่อน +4

    Great video as always. Keep going dude!

  • @HarpaAI
    @HarpaAI 3 หลายเดือนก่อน

    🎯 Key Takeaways for quick navigation:
    00:00 🧠 Alpha Geometry Achievements and Implications
    - Alpha Geometry achieves high scores in the International Math Olympiad (IMO) for geometry problems.
    - The system combines neural networks with symbolic pre-programmed systems, showcasing the alliance between language models and traditional problem-solving methods.
    - Despite not achieving AGI, advancements in neuro-symbolic systems like Alpha Geometry suggest progress towards more sophisticated AI models.
    03:12 🔍 Mechanism of Alpha Geometry
    - Alpha Geometry's approach involves a neural language model proposing solutions combined with a symbolic engine for mechanical deduction.
    - The system utilizes synthetic data for training and can generate novel mathematical constructs to solve problems.
    - It iteratively refines solutions through a loop of proposing constructs, solving, and adjusting based on feedback.
    05:20 🔄 Neuro-Symbolic Approach and Generalization
    - The neuro-symbolic approach of Alpha Geometry isn't entirely novel but represents a significant advancement in solving mathematical problems.
    - Generalizing such systems across mathematical fields is a current focus, with potential applications beyond mathematics.
    - While achieving impressive results, these systems still lack the symmetrical elegance of human-discovered theorems but focus more on functionality and effectiveness.
    07:21 💻 Hardware and Future Prospects
    - Alpha Geometry's performance could further improve with advancements in hardware, notably with GPUs like Nvidia's V100.
    - The evolving hardware landscape suggests continued progress in AI capabilities, potentially leading to solving more complex problems in the future.
    - Discussions about the implications of AI advancements, including reasoning capabilities and the potential for AGI, remain ongoing.
    09:38 🌟 Alpha Codium and Iterative Problem-Solving
    - Alpha Codium, an open-source project, demonstrates a similar theme of iterative problem-solving through LLMs proposing solutions and adapting based on feedback.
    - The process involves reasoning about problems, generating potential solutions, and refining them through testing and iteration.
    - This approach contrasts with the traditional prompt-to-answer method, showing a shift towards more interactive and conversational interactions with LLMs.
    Made with HARPA AI

  • @vladyslavkorenyak872
    @vladyslavkorenyak872 4 หลายเดือนก่อน +3

    There are many algebraic problems that can be transformed into geometric problems. I predict there will be future developments that integrate all of math.

  • @user-hk8jt6so3l
    @user-hk8jt6so3l 4 หลายเดือนก่อน +1

    Thank you so much! I wish I could explain how happy I get when I see you releasing a new video, you rule since day 1!❤
    edit: I've decided, I NEED to get access to your patreon no matter what!

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน

      Thanks so much my man. Feel free to dip in and out!

  • @Bonirin
    @Bonirin 4 หลายเดือนก่อน +7

    I actually saw this guy on IG several times, and although I would say that explaining math using celebrities COULD be beneficial, I think this guy is just good at explaining math

    • @ChristopherLegend
      @ChristopherLegend 4 หลายเดือนก่อน

      What’s his ig handle? Thx

    • @Bonirin
      @Bonirin 4 หลายเดือนก่อน

      @@ChristopherLegend @onlocklearning

  • @memegazer
    @memegazer 4 หลายเดือนก่อน +1

    Was waiting for this vid and your insights did not disapoint...love you bro...thanks for keeping us informed

    • @memegazer
      @memegazer 4 หลายเดือนก่อน

      I really feel like sometimes optimists do not appreciate problems surrounding search and storage.
      Examples like this are a great way to approach the type of opptimism expressed by the likes of Wolfram for example.

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน

      Thanks meme, as always

  • @avidlearner8117
    @avidlearner8117 4 หลายเดือนก่อน

    The second part is MASSIVE. It shows how much there is still possibilities withing GPT-4, just by proper prompt engineering.... Fantastic!

  • @stephenrodwell
    @stephenrodwell 4 หลายเดือนก่อน +1

    Thanks as always! Brilliant content. 🙏🏼

  • @ObservingBeauty
    @ObservingBeauty 4 หลายเดือนก่อน +2

    So much value packed in one video. Thanks

  • @anywallsocket
    @anywallsocket 4 หลายเดือนก่อน +4

    All of science can be crudely summarized by the guess and check iteration method, so why not AI?

  • @AllisterVinris
    @AllisterVinris 4 หลายเดือนก่อน +1

    The pieces are getting in place. Who knows how long we have before the puzzle is finally complete?
    AGI hype

  • @christophchristoph9909
    @christophchristoph9909 4 หลายเดือนก่อน

    When you says give the model space to explore (e.g. through Reflexion) its important to remember that this is only successful with the super powerful models such as GPT-4 or GPT-4 Turbo. With GPT 3.5 Turbo Reflexion and chain of thought is super inconsistent and at times arbitrary. This is why I like the Test Time approach - which has been proven to upscale the performance of smaller models. Also, a verifier model ranking according to scalar estimates such as ‚estimate the revenue this business plan will make in €‘ is a super straight forward scoring system which makes it super easy to filter out ideal solutions among thousands of suggestions, i.e. it can be scaled easier and implement in a straight forward way.

  • @HoD999x
    @HoD999x 4 หลายเดือนก่อน +1

    did you read my comment about the "haitch"? :D
    i was wondering if this was a similar thing as "RTX 3080 Tee Eye" vs "Tie" where nvidia itself is inconsistent

  • @brianWreaves
    @brianWreaves 4 หลายเดือนก่อน +1

    Another brilliant share! 🏆

  • @amusik7
    @amusik7 4 หลายเดือนก่อน

    I can also imagine that the most powerful AIs will integrate these engines as modules (like AlphaGeometry) they can call upon when needed. So you will have some particularly powerful GPT organising and understanding and pulling in the resources as needed.

  • @DreckbobBratpfanne
    @DreckbobBratpfanne 4 หลายเดือนก่อน +1

    The biggest accelerator is probably gonna be continuous learning... Because assume you use such a structure, as seen in these examples, but the LLM that does the work is improving as well with each step. The more it fails the better it gets, the more it solves the even better it gets. Until the limit for its size are reached (and now assume you could introduce new neurons after that point to top it off)
    I wonder when the first real world system uses this structure, for example a cloud lab for chemistry

    • @colejohnson2230
      @colejohnson2230 4 หลายเดือนก่อน +1

      That's what I'm waiting for, and I agree wholeheartedly.
      I'd imagine we will get something like that which has a sort of accelerated growth until it hits a sort of compute/resource ceiling from allocating new neurons/connections or whatever. It's not hard to see after that a quick adjustment that allows it to cleverly forget things that aren't important, which is exactly the sort of breakthrough that brought us the attention mechanism, just focusing on the opposite thing. Who knows what happens then

  • @roelzylstra
    @roelzylstra 3 หลายเดือนก่อน +2

    Thank you for referring to Twitter as Twitter.

  • @nuclear_AI
    @nuclear_AI 4 หลายเดือนก่อน +4

    The future of education is mind boggling! The engineers of tomorrow are going to have fun jobs.
    Our children are going to build things we lack the imagination to even conceive 🤯🧠

  • @chinchao
    @chinchao 4 หลายเดือนก่อน +1

    Brilliant content, you should have a podcast 😊

  • @jsx00
    @jsx00 4 หลายเดือนก่อน +1

    Honey wake up, new AI Explained video just dropped

  • @mattmexor2882
    @mattmexor2882 4 หลายเดือนก่อน +1

    Mathematics isn't just about proving things. It's about making abstractions and recognizing patterns and associations. In order for mathematics to "fall" an AI would have to at least develop conjectures and prove them (it should probably also be able to develop a construction that could be usable to study a problem of interest). Making sensible conjectures may demand AGI. And the search space for developing new constructions, whether to make associations between different areas of knowledge or just to prove a conjecture, is likely way too large without judgment. But a proof bot could be a tool that a mathematician uses to explore mathematics.
    It should also be noted that it sounds like this bot was trained on proofs in a well-developed area of mathematics. If I understood, the training data was other proofs in that area. Using well-developed methods in new situations that closely follows the original use is certainly very useful in mathematics, but even were the bot to be used only to find proofs as directed by a human, and nothing more, if the bot were to lose effectiveness in a novel area of math where new axioms and definitions were introduced it would greatly reduces its usefulness, especially if it's even unable to abstract known methods from other areas of math and apply them to the novel area.
    Most people just don't understand what mathematics entails. The fact that math can be parsed easily and uses a very precise language is great for an AI. But mathematics requires a tremendous amount of creativity, ability to abstract, and ability to reason. Symbolic reasoning is not enough, there needs to be deeper reasoning. You can get a whole lot done with effective symbol-pushing. You can get a whole lot done with intelligently searching through a database of proof ideas. You can get even more done when you combine the two together. But you can't crack mathematics with it. You're not even close.

  • @YoussefMohamed-er6zy
    @YoussefMohamed-er6zy 4 หลายเดือนก่อน +2

    Told ya we need Symbolic systems still

  • @GilesBathgate
    @GilesBathgate 4 หลายเดือนก่อน

    Backtracking: What if we can train an LLM with, "This is simple sentence that illustrates jahskd correcting previous tokens." i.e The training data is augmented by just inserting a noisy token followed by a special backspace token. At inference, the special token removes the noise and backspace from the context window on just that itteration.
    Extend the idea, and you now have an LLM that can re-think things starting from an arbitrary point in the past?

  • @Rick_Mactavish
    @Rick_Mactavish 4 หลายเดือนก่อน +1

    Thanks for the hard work!

  • @antman7673
    @antman7673 4 หลายเดือนก่อน +1

    Winning a math Olympiad is very significant.
    When you can not just replace a bad worker, but the best worker, uuuhhh weee.

  • @jtangen
    @jtangen 4 หลายเดือนก่อน +3

    Great video! Can you please add a link to the Kardashian teacher video?

  • @rrioclkls7721
    @rrioclkls7721 4 หลายเดือนก่อน

    It's interesting, LLMs and ML in general are created by essentially bruteforcing weights & testing until it gives you the answer you want, what alphageometry, codium & eureka do use LLMs to get a good guess as to where to start instead of from scratch. Its like one off last mile training

  • @federicoaschieri
    @federicoaschieri 4 หลายเดือนก่อน +2

    Interesting paper, but if they said to not overhype, there is a reason. The euclidean geometry considered in this paper is a so-called *decidable* theory, which means that there is actually an algorithm that mechanically solve *all* problems in a finite amount of time: it is the algebra-based one. They just didn't let the algebraic algorithm run for more than 48h 😁 So this kind of problems allow brute force, because no intelligence is actually required to find a solution. It is a step to AGI as Alpha-go is: none. This is an example of narrow AI that can solve problems in domains that just require "smart brute force". Alas, true mathematics is a different beast. It is undecidable: there cannot be any algorithm that solve all problems. So the methods in this paper won't really scale, not at human-level. They could provide just useful proof-assistants. My prediction of no AGI in foreseeable future holds true once again.

  • @reason239
    @reason239 4 หลายเดือนก่อน +1

    At 8:44 you say that Google plans to add AlphaGeometry to Gemini, but you show the post that seems to talk about adding the visual input to AlphaGeometry using Gemini. Aren't those different?

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน

      As in it might be a yet improved version that gets in

    • @reason239
      @reason239 4 หลายเดือนก่อน

      @@aiexplained-official ah, ok, thanks for the clarification. Could you please share the source that promises to add AlphaGeometry to Gemini?
      I'm interested because I'm an IMO participant myself, and my math teacher is worried that AlphaGeometry could be really detrimental to Math education. Because if this tool becomes easily available, this will make students less motivated (because it's so easy to cheat and ask the model to solve the problem instead of solving it yourself)
      I guess the AI has already kind of ruined the home assignments like essays.
      Do you have an opinion on whether AI will have a net positive or negative impact on education?
      I could see arguments for both. I'd guess the education system will definitely have to adapt.

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน +1

      Hey reason, all sources in the description, think it was the NYT one. And well done on getting to IMO, further than I ever did. As for education, honestly I feel, by the 2030s, it will be much more like edutainment for most people, even schools. IMO participants might even be like chess superstars, where everyone can watch along while an AI explains what they are attempting.

  • @NobleCaveman
    @NobleCaveman 4 หลายเดือนก่อน +1

    Excellent video

  • @lthedoperabbitl9258
    @lthedoperabbitl9258 4 หลายเดือนก่อน +1

    will we ever get those videos on the main channel ? like maybe later? :( i cant afford it otherwise i would have 100% been joined insiders :(

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน

      I understand ! One option is to sub after a whilr and then binge everything! Then unsub after. Many people can write it off as training expense too, if you are employed.

  • @Cygx
    @Cygx 4 หลายเดือนก่อน +1

    These large language models is actually reflective of progress and development happens in the real world. First principal allows us to not pollute the solution space and these AI models seem to be very good at suggesting novel solutions.

  • @ellielikesmath
    @ellielikesmath 4 หลายเดือนก่อน

    i cant help but notice i did sort of call (in a comment on this very channel) that using auto deduction/formal inference to generate synthetic data would help. not necessarily hard to call, but still very nice and quite surprising as well to see it happen so quickly! on the other hand, we saw gpt-4 solve imo problems before, so it wasn't entirely new. i thought the approach they used could probably be refined (a lot). for instance, the synthetic data generation step ought to be tried to be filtered for "better" proofs (shorter, more generalization, more symmetry, etc.). the idea that this would represent some "aesthetic" bias is absurd. those kinds of "biases" help human math people because, as it turns out, on any number of fundamental levels, nature looks symmetric. also, the more mystical and magical ai looks, the less progress we will make in thinking of ways to improve it (not that we are in any danger of that!!). thanks for the video👍

  • @jonp3674
    @jonp3674 4 หลายเดือนก่อน +1

    Why is the estimate for gold in the IMO 2027? Is there a reason why this exact approach wouldn't work for the other questions?
    I think this "intuition-er" + "reasoner" back and forth will work for the other areas of the test and there's about an 80% chance gold in the IMO will fall in 6 months time.

  • @MrMatklug
    @MrMatklug 4 หลายเดือนก่อน +1

    will we get life extension explained?, in a older video you said that liked the accompany the news

  • @DailyOverdoseofPapers
    @DailyOverdoseofPapers 4 หลายเดือนก่อน

    I'm curious would using more principled reasoning mechanisms from RL make this even better?

  • @HERKELMERKEL
    @HERKELMERKEL 2 หลายเดือนก่อน

    btw, if you draw the given data correctly on geogebra, it automatically shows the result.. because geometry problems (intependent from its difficulty) are made with at least enough or more information on angles or lenghts etc (the data), in order to secure that there is only one solution to that particular problem.. so sometimes it can be hard to solve for humans to see.. but it cannot be the case for machine. it outta be precise.. it easily measure the distance or angeles etc and show us the result.. but solving with rules, theorems, lemmas, axioms, trigonometric tabels etc. i mean solving without measuring is another thing.. in that case, the AI comes in.. but all my tryouts via gemini, chatgpt, copilot.. i didnt get right answers on math or geo problems.. i hope it will change in the near future... hopes up on alphageometry

  • @melvingeraldsy1552
    @melvingeraldsy1552 4 หลายเดือนก่อน +1

    Tech ceos are now always talking about agi in X, they are probably easing the public for its arrival. Looks like it really is coming very very soon.

  • @musicarroll
    @musicarroll 4 หลายเดือนก่อน

    Santiago's steps are also eerily similar to standard TDD, Test Driven Development.

  • @user-if1ly5sn5f
    @user-if1ly5sn5f 4 หลายเดือนก่อน

    4:58 so the ai has an ooda loop now. Sweet so it’s better than before. Plus if we speed up this looping process and add more then we basically are backwards engineering human through a different medium kinda like how a copy is a portion but then put the portions together and you have the real thing just separate from the current one.

  • @ginogarcia8730
    @ginogarcia8730 4 หลายเดือนก่อน +1

    another 5 years of AI significance worth jampacked in one AI explained video

  • @andersonsystem2
    @andersonsystem2 4 หลายเดือนก่อน +1

    Good video mate 🎉

  • @hydrogenbond7303
    @hydrogenbond7303 4 หลายเดือนก่อน +1

    Wow just that short clip finally made me understand integrals 😂😂😂
    No one ever explained It to me like that 😂

  • @OneSlavBoi
    @OneSlavBoi 4 หลายเดือนก่อน +1

    okay but how do you do all these videos and prepwork, it is genuinely easy to be in disbelief since you do all this alone.

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน +1

      Yeah so far everything, the channel, Insiders, all alone

  • @Xilefx7
    @Xilefx7 4 หลายเดือนก่อน +1

    Good video as always

  • @wale7342
    @wale7342 4 หลายเดือนก่อน +1

    Baby wake up, ai explained just dropped a video

  • @NinitoPh
    @NinitoPh 4 หลายเดือนก่อน +2

    Awesome time to be subscribed to AI insiders! You absolutely carpet bombed us with content today.

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน +1

      Amen! 14 hr day for sure, based on weeks of previous research.

  • @christophschwab-ganser9366
    @christophschwab-ganser9366 2 หลายเดือนก่อน +1

    Thanks!

  • @ahrenadams
    @ahrenadams 4 หลายเดือนก่อน

    It seems Hallucination can be used as a benefit for finding novel solutions that can be tested and thrown away if not applicable just like a human would have epiphanies and try and test said epiphany?

  • @agush22
    @agush22 4 หลายเดือนก่อน

    Would you consider some discount on the insiders for people living in developing economies?

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน

      I wish I coupd but Patreon doesn't allow this, unless you can find a way that I can't. Don't even allow promo codes.

    • @agush22
      @agush22 4 หลายเดือนก่อน

      @@aiexplained-official thanks for the reply and for the amazing content, cheers!

  • @Robert_McGarry_Poems
    @Robert_McGarry_Poems 4 หลายเดือนก่อน +1

    Crazy times we live in.

  • @DanielSeacrest
    @DanielSeacrest 4 หลายเดือนก่อน +2

    Meta just announced they are aiming to have the equivilent compute of 600k H100s by the end of the year. That is around 2.4 zettaflops AI compute power (FP8) or 1.2 zettaflops with FP16. With 3x gain from good data quality, Active learning provides another ~3x gain and 3x gain and architecture improvements (RoPE, SwiGLU ...), and using FP8 with 50% utilisation we could see a model trained with 11,718x effective compute over GPT-4 in 90 days at the start of 2025😂💀 (assuming 2.15x10^25 FLOPs for GPT-4s training). Also i could see upwards of 10x gain from data quality which would give 23,437x effective compute over GPT-4 assuming about 50% utilisation. If that isn't an exponential cureve i don't know what is lol.

  • @KevinD7
    @KevinD7 4 หลายเดือนก่อน +2

    very interesting!!

  • @kennethmyers6160
    @kennethmyers6160 4 หลายเดือนก่อน +2

    Llms are the right brain of AGI

  • @CaspersCuts
    @CaspersCuts 4 หลายเดือนก่อน +1

    TIL I pronounce my Hs in a Cockney way. Must call the Trouble and Strife on the old Dog and Bone to let her know.

  • @RonLWilson
    @RonLWilson 4 หลายเดือนก่อน +1

    This looks very promising!

    • @RonLWilson
      @RonLWilson 4 หลายเดือนก่อน

      One approach might be applied to not just geometry algorithms but other standard standard algorithms as well such as Dijkstra's algorithm (such as used for route planning) and Auction algorithm (the resource assignment problem) as well with dynamic program algorithms and game theory.
      Also, the AI might help with the costing of these.

    • @RonLWilson
      @RonLWilson 4 หลายเดือนก่อน +1

      AI seems to serve much as human intuition were more standard algorithms are more logic based where the two can work together.

  • @neelmehta9092
    @neelmehta9092 4 หลายเดือนก่อน +1

    I find the exponential curve quite worrisome, in 2 years we have cracked geometry, and this too with models which are reportedly, by sam altman of all people, supposed to be "very bad" at reasoning. They didn't even use the H100s for this, not only this but meta training an open source GPT-4 competitor just doesn't sit right with me. I am all for opensourcing but we have only had GPT-4 for about a year now and we are still only just figuring out its abilities to crack geometry. What if in 2 months a prompt engineering technique comes along that makes GPT-4 really good at nuclear science? We don't know what the ceiling of this model is and until we can find that, open-sourcing such powerful models is risky

  • @TheVineOfChristLives
    @TheVineOfChristLives 4 หลายเดือนก่อน

    Where do I find that math stuff? I have kids that I need to have that kind of help with.

  • @ryzikx
    @ryzikx 4 หลายเดือนก่อน +4

    "self play" on pure mathematics and code, which has clear objective functions. i can't wait to see more of this.

  • @bobtivnan
    @bobtivnan 4 หลายเดือนก่อน +1

    Is there a link for the AI generated calculus teachers?

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน

      In description

    • @bobtivnan
      @bobtivnan 4 หลายเดือนก่อน

      @@aiexplained-official ty. I missed the ai Kardashian link at first. I am skeptical of this. It feels too polished as though there was lots of post production editing to make it look like a natural interaction. This is probably not too far away though. As a math teacher, the casual conversation design intrigues me- this is how I engage students. These AI Kardashians remind me of the Nancy Pi videos on TH-cam. I would love to see an AI teacher that uses her Lightboard style of math explanation.

  • @novadea1643
    @novadea1643 4 หลายเดือนก่อน +1

    I don't think a lot of the machine optimized geometries/solutions etc. look like trash, but they do look alien for sure, something no human would or could design. They often also seem/are more "natural/biological" because apparently evolution having optimized structures for billions of years has gotten quite a bit of it, if not right, close enough to us to want to mimic for next step up in technology.

  • @bluetee.531
    @bluetee.531 4 หลายเดือนก่อน

    Bro. Can you start a Twitch Channel and stream reading papers and giving your assessments?

  • @davidebic
    @davidebic 4 หลายเดือนก่อน

    I mean, they just have a neural network whose purpose is to hallucinate some steps that are commonly used... What makes it a viable strategy is just the mechanical solver. But it makes for such a great idea to implement in a chess engine. A neural network that first hallucinates a sensible move to make the alpha-beta pruning faster.

  • @Theonlyrealcornpop
    @Theonlyrealcornpop 4 หลายเดือนก่อน

    Phillip, I think the thing I want to know is whether or not you think the development of AlphaGeometry is proof of the validity of what that one Chinese guy said in the interview in AI Insiders a couple weeks ago where he said something akin to "Functions (or tokens, I forget which) are all you need", since it seems to me, anyway, that LLM -> Solver loop demonstrates that his intuition (or, as he said, YOUR intuition) within that statement really is beginning to open up a whole new class of L(X)M innovations that are so much more sophisticated in their end use cases

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน

      It's definitely linked and could be expanded with time, for sure

  • @SickPillow
    @SickPillow 4 หลายเดือนก่อน +3

    Drinking game: Drink whenever he says LLM. (I died)

  • @karthage3637
    @karthage3637 4 หลายเดือนก่อน +2

    nice i needed something to watch

    • @aiexplained-official
      @aiexplained-official  4 หลายเดือนก่อน +1

      Let me know what you think!

    • @karthage3637
      @karthage3637 4 หลายเดือนก่อน

      @@aiexplained-official I am quite stun by the ability of deep mind to produce new expert AI on various theoretical field. It also make me wonder why gemini feels a bit of compare to the deepmind standard. It felt like it was push out with the deepmind brand a bit early in the process. But with what we learn from the secret GPT4 sauce and Mixtral. I expect gemini to really shine once all of this army of expert will be gathered under one maestro to work in concert. Wait and see.

  • @phillaysheo8
    @phillaysheo8 4 หลายเดือนก่อน +2

    Thanks Dude. 2024 is going to be even more exciting than 2023 it seems.

  • @invizii2645
    @invizii2645 4 หลายเดือนก่อน +1

    Nice

  • @Skwiggs
    @Skwiggs 4 หลายเดือนก่อน

    Every single one of your videos makes me increasingly excited for the future

  • @GearForTheYear
    @GearForTheYear 4 หลายเดือนก่อน

    8:30 The ability to reason is not a magical concept -- it is grounded in the relationships between symbolic representations of concepts. A task perfectly suited to LLMs. I don't really get why there's still so much 'debate' about this. The only reason why public-facing LLMs appear to have trouble reasoning is because they aren't given space to reason before providing an output due to compute costs.

  • @JazevoAudiosurf
    @JazevoAudiosurf 4 หลายเดือนก่อน +2

    i read alphacopium

  • @user-if1ly5sn5f
    @user-if1ly5sn5f 4 หลายเดือนก่อน +1

    0:50 yeah ai is the way to go in learning. Honestly the current schools are bad and that’s why the wealthy have private schools or home school with expensive tutors. Now we could have the smartest people in ai form and not only teach but listen to the kids or adults and learn the best teaching methods for them and even understand if they understand the processes and not just the information. Children are hard to teach so the ai could use differences to teach them easier, like with cartoons or interactive games or even books and interactive books. The ai could be a companion that goes everywhere and helps with anything you have trouble with. A ghost from destiny is a good example. Schools may be a thing of the past if we start making ai companions with teaching functions like a twin that walks beside your life and learns too about you. This could be a symbiotic relationship.

  • @anatolwegner9096
    @anatolwegner9096 4 หลายเดือนก่อน +1

    It's funny how much the presenters voice resembles D. Hassabis.

  • @trentondambrowitz1746
    @trentondambrowitz1746 4 หลายเดือนก่อน +1

    3 in one day!

  • @AngeloXification
    @AngeloXification 4 หลายเดือนก่อน +1

    10:56 that's hilarious that's exactly how I code with chatgpt

  • @oo__ee
    @oo__ee 4 หลายเดือนก่อน

    If it's going to solve all geometry problems, my question becomes "aren't all problems geometry problems?"

  • @IntentStore
    @IntentStore 4 หลายเดือนก่อน

    The better general purpose ai we get is going to be much more useful than these super overfit boutique solutions to specific problems

  • @JazevoAudiosurf
    @JazevoAudiosurf 4 หลายเดือนก่อน +2

    you know i didnt get into harvard, i didnt study that hard. loool

  • @sidnath7336
    @sidnath7336 4 หลายเดือนก่อน

    This is an incredible project but, to me, it could be more remarkable.
    Imagine the human brain, which is somewhat already adept in Mathematics, being able to practice 100 million geometry problems and being able to retain each method and solution - we would perhaps see the same level of “scaling”, particularly because human cognition abilities, in terms of contextual and intuitive reasoning, is still somewhat superior to AI and also explainable.
    What would make this remarkable to me is if we had a way to test the sample complexity, in terms of:
    i) Finding the bare minimum amount of training examples needed to reach SOTA level.
    ii) The exact examples in which the AI “abstracts” its answers from when approaching a new question i.e. can the AI solve a difficult geometry problem from a collection of unrelated geometry problems or is it genuine few-shot learning?
    There are many more things that can be researched, especially the explainability part, but nevertheless this is a fantastic step towards enhanced reasoning and problem solving capabilities.
    DeepMind all the way! 💯