AI just got memory & learning - Google's INSANE breakthrough

แชร์
ฝัง
  • เผยแพร่เมื่อ 1 ก.พ. 2025

ความคิดเห็น • 314

  • @iweather-nr6kp
    @iweather-nr6kp 10 วันที่ผ่านมา +347

    "The Singularity" = AI videos every 5 milliseconds

    • @ZorenStudio55
      @ZorenStudio55 10 วันที่ผ่านมา +17

      IMG to reality 😂😂
      Like , you upload a woman img , write prompt and you would get that 😂😂😂

    • @David0gden
      @David0gden 10 วันที่ผ่านมา +12

      ​@@ZorenStudio55
      All I know is if we get carbon printers that would be capable of making objects or lifeforms/AI exoskeletons... HP is going to take all our money for the cartridges.

    • @kuromiLayfe
      @kuromiLayfe 10 วันที่ผ่านมา +6

      the AI is already capable of doing that.. it is just that our hardware is unable to render the output at that speed… due to bottlenecks.

    • @House-Metal-Punk-And-Your-Mom
      @House-Metal-Punk-And-Your-Mom 10 วันที่ผ่านมา

      video clones every seconds

    • @ZorenStudio55
      @ZorenStudio55 10 วันที่ผ่านมา +2

      @@David0gden future is here

  • @ShifterBo1
    @ShifterBo1 10 วันที่ผ่านมา +94

    40m long video from "AI Search" isnt something i was expecting on my bingo card

  • @LukelayAlt
    @LukelayAlt 10 วันที่ผ่านมา +89

    Meta's Large Concept Models for thought abstraction
    Google's Titans for human-like memory
    Sakana's Transformers² for more human-like Mixture of Experts
    DeepSeek's Reinforcement Learning on LLMs for automatic self-learning
    2025 sure is gonna be interesting huh--

    • @dascodraws6040
      @dascodraws6040 10 วันที่ผ่านมา +11

      the infinity stones or forming captain planet

    • @wilburdemitel8468
      @wilburdemitel8468 10 วันที่ผ่านมา +1

      @dascodraws6040 yes, it has about the same real world usage as those fictional concepts

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา +14

      and it's only Jan

    • @limitationsoflanguage
      @limitationsoflanguage 10 วันที่ผ่านมา +2

      How do we layer all these different architectures' functionality into a single point of user interaction modularly?

    • @tiergeist2639
      @tiergeist2639 10 วันที่ผ่านมา +5

      ​@@limitationsoflanguagethats the smallest problem. building connections is not too difficult.

  • @llElektrobeat
    @llElektrobeat 7 วันที่ผ่านมา +1

    Thats the first video i found from this YT channel and i never meet some one who really understands the AI topic so deeply and can explain it so well.
    Cant tell you, how much stuff i have learned through this single video. Looking so hard forward for your other videos!
    Insane work!

  • @sidcoolguy7
    @sidcoolguy7 10 วันที่ผ่านมา +16

    Timestamps (Powered by Merlin AI)
    00:05 - Google's Titans AI architecture significantly advances memory and learning capabilities.
    02:28 - Google's Titans architecture addresses the limitations of existing Transformer models.
    07:07 - AI models can learn and adapt during test time with Titans technology.
    09:17 - AI incorporates long-term memory inspired by human learning mechanisms.
    13:57 - Google's TITANS architecture introduces advanced memory mechanisms for AI.
    16:17 - AI utilizes different memory types for enhanced information processing.
    20:26 - AI advances in complex data analysis but faces slower training times.
    22:32 - New AI memory mechanisms improve context understanding for conversational tasks.
    26:50 - Google's Titans paper signals a new wave of evolving AI models.
    28:57 - AI evolves like the human brain through neuroplasticity, adapting in real time.
    33:10 - AI classifies tasks using three adaptation methods.
    35:20 - Transformer Squared models adapt in real time to enhance AI performance.
    39:41 - Emerging AI models may surpass current Transformer capabilities.

  • @IceMetalPunk
    @IceMetalPunk 10 วันที่ผ่านมา +22

    I want to point out two things about the Sakana T² architecure:
    1. From my understanding, it's not really continual learning: the updated weights aren't stored or overwritten beyond the current prompt. It just boosts and suppresses different neural cliques for the individual prompt and not longer, making it more like a "task focus" than continual learning.
    2. It can be applied to any Transformer model. While Titans is an iteration on existing architectures, the core layer of its memory-as-context variant is very similar to a Transformer. Which makes me wonder if it would be relatively simple to combine T² with Titans and get actual continual learning *plus* the boost of task focus?

    • @automatescellulaires8543
      @automatescellulaires8543 10 วันที่ผ่านมา

      i'm not sure the author care about such details though. Nor does the audience i believe. Thank you for pointing it out, for guys like. Though i'll try to make youtube not recommend this type of channel in the future. It feels like the author doesn't have a very clear picture of what he is talking about.

    • @IceMetalPunk
      @IceMetalPunk 10 วันที่ผ่านมา +6

      @@automatescellulaires8543 That's a bit harsh. People make mistakes, I don't assume apathy or malice from the start.

    • @automatescellulaires8543
      @automatescellulaires8543 10 วันที่ผ่านมา +2

      @@IceMetalPunk You are a good human. i'm not. The channel name, and the narration makes one believe the thing is leggit, when it clearly is not. Maybe it's a young youtuber that will build competency after a while, but i don't have the time to be mislead like that. It's not the worst out there, but still pisses me.

    • @Codemanlex
      @Codemanlex 10 วันที่ผ่านมา

      What a time to be alive right?
      I actually thought of this when I read the papers weeks ago.
      This is why I always say open source is the best because we all can help build better systems.

    • @Codemanlex
      @Codemanlex 10 วันที่ผ่านมา +2

      @@automatescellulaires8543Why do you say such words?
      This guy has a good channel and helps lots of people understand things like this.
      Maybe start your TH-cam channel instead?
      Seems you know better
      I would also like to learn more too

  • @frotercannon2740
    @frotercannon2740 8 วันที่ผ่านมา +5

    Now all that needs to happen is Titan and transformer^2 to come together so by memory it can compare efficiency of the answer and alter the structure of the nodes, weights, and balances accordingly. Basically, an artificial soul.

  • @sylvaindelaine
    @sylvaindelaine 10 วันที่ผ่านมา +9

    Thank you for providing such valuable information and for demystifying AI, along with the latest updates and models. Your clear and accessible explanations have been incredibly helpful, and I am learning a great deal from them. I truly appreciate it.

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา

      You're welcome!

  • @danielle78730
    @danielle78730 10 วันที่ผ่านมา +4

    unlike many AI vids these days, i found your visual content every bit as helpful in conveying your content as was your script. (i tried listening passively from another room and decided to restart the entire video whilst giving it my full attention and squinting at the screen.) in other words, as always, thank you! your detailed research and refusal to dumb things down is *much* appreciated!

  • @FerdinandCesarano
    @FerdinandCesarano 10 วันที่ผ่านมา +6

    31:40 - Note that, in "occipital", the first C is hard and the second C is soft. So the word is pronounced as though it were spelt "oksipital".

  • @nehemiasvasquez8536
    @nehemiasvasquez8536 10 วันที่ผ่านมา +3

    It is very interesting how these Titans can gather memory out of new events triggering them. If they add the stategies used on the Transformer² for making the Titans more flexible, it would become much more its own living and evolving thing, eventually.

  • @AndyBarbosa96
    @AndyBarbosa96 9 วันที่ผ่านมา +3

    Brilliant summary, definitely the future, neuroplasticity in both content and form: weights and architecture. The new generation of AI architectures will definitely allow for the continuous adaptation of their modular constituents combined with dynamic memory. Being also able to pick an existing model from scratch and 'adapt' ir enable it with such features is brilliant, since this will allow for architectural research built upon foundational work, opening this radical new space to new joiners with great ideas but no capital nor millions of gpus available. Coincidentally, this is the core of our own research, and these 2 papers are very much seminal/topical in this space. We call it TnT transformers to the nth, since our architectures are multidimensional tensors of neurons and not flat 2d layers. So the tensor Algebra is definitely more complex, but the idea is very much the same. Thanks for your amazing work. We will be using distilled DeepSeek r1s as prototypes since these open source open weights.

  • @Darknight_84930
    @Darknight_84930 10 วันที่ผ่านมา +23

    Honestly this maybe the best time to start your life with all these ai breakthroughs like me going through my teenage life

    • @GraipVine
      @GraipVine 10 วันที่ผ่านมา +6

      I'm just glad to see the beginning. I'm very old so won't see the end, or perhaps even the middle. But the start of something new, different, and groundbreaking is always exciting!

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา +11

      I'm glad I'm alive to see this. We are at a pivotal time in history.

    • @limitationsoflanguage
      @limitationsoflanguage 10 วันที่ผ่านมา

      You only got ten years before we're all driving flying Lamborghinis around the moon! (We will have self-driving technology in space by then but it won't be allowed for another 20 years after due to safety concerns) th-cam.com/video/8TH3gvdaK18/w-d-xo.htmlsi=xTB3gf1mnPNomGQ8

    • @aguspuig6615
      @aguspuig6615 10 วันที่ผ่านมา +4

      @@GraipVine Hey you never know, with how AI is beating even the optimistic predictions who knows if we will get advancements in health that keep people alive... well god knows for how long. Hopefully it happens, and fast enough for you brother

    • @aguspuig6615
      @aguspuig6615 10 วันที่ผ่านมา +4

      Tbh, kids born now will probably not even learn the concept of work as we know it. Hell im studying in college rn and i dont know if ill even have to work by the time im done

  • @GraipVine
    @GraipVine 10 วันที่ผ่านมา +13

    About time. I've been emailing them on a steady basis with these designs based on modular reasoning and using subsets for particular requirements, as well as breaking down the structure of memory for them to adapt to AI, attention, long-term, persistent, as performed by the human brain. I'm glad they're now doing this process. (Yes, I have heaps of conversations with LLMs involving these concepts from months ago.) Finally, they've listened!

    • @jawbone3910
      @jawbone3910 10 วันที่ผ่านมา +25

      I’m sure that your emails were definitely what led the genius AI researchers at Google to develop Titan

    • @201crunt6
      @201crunt6 10 วันที่ผ่านมา +7

      Send a invoice

    • @TragicGFuel
      @TragicGFuel 10 วันที่ผ่านมา +6

      So you basically talked to an LLM and then emailed google to tell them how they should research?-

    • @kulyog6626
      @kulyog6626 10 วันที่ผ่านมา +1

      Hi

    • @GraipVine
      @GraipVine 10 วันที่ผ่านมา +1

      @@TragicGFuel The same way they use AI to synthesize data and come up with solutions humans would not have dreamed up? Well, duh, yes.

  • @JustChillinOnThe5thFloor
    @JustChillinOnThe5thFloor 10 วันที่ผ่านมา +4

    "This changes everything" is the moment AI Search tells us he was an AI this time

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา +2

      I'll say this on 4/1 😉

  • @WiseApe9
    @WiseApe9 10 วันที่ผ่านมา

    Thanks for the vid. I appreciate the longform content

  • @ZorenStudio55
    @ZorenStudio55 10 วันที่ผ่านมา +15

    Ai never sleeps,
    Now ai is evolving, right?

    • @dtrueg
      @dtrueg 10 วันที่ผ่านมา +1

      no

    • @janebajWa
      @janebajWa 10 วันที่ผ่านมา +4

      sleep when no electricity

    • @MAGNETO-i1i
      @MAGNETO-i1i 10 วันที่ผ่านมา +2

      ​@@janebajWaUse humans as batteries

  • @westwest7349
    @westwest7349 8 วันที่ผ่านมา +5

    Being a pro in AI for last 15 years, I can tell it is an absolute huge step forward.

  • @am243pm
    @am243pm 10 วันที่ผ่านมา +2

    Just wanted to say nice video! Thank you for taking the time to digest this information into a easily understandable format!! Will definitely be staying tune to more of your breakdowns on publications 💪💪

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา +1

      You're welcome!

  • @MrJohnnyseven
    @MrJohnnyseven 10 วันที่ผ่านมา +16

    Everyday there is 10 or more videos saying "it will change everything" lol

    • @MrJohnnyseven
      @MrJohnnyseven 10 วันที่ผ่านมา +4

      And of course it's gonna be "insane"...😂

    • @MAGNETO-i1i
      @MAGNETO-i1i 10 วันที่ผ่านมา +4

      Meanwhile:
      "Sorry for any confusion, but as an AI..."

  • @verlax8956
    @verlax8956 10 วันที่ผ่านมา +7

    what's the difference between this and liquid neural networks if they can both actively learn?

    • @bennyboiii1196
      @bennyboiii1196 10 วันที่ผ่านมา +13

      As far as I understand, the LLM (core), the Knowledge Base (Persistent Memory) and the Memory (Long Term Memory) are three separate MLP's that reference each other. The actual LLM doesn't change at inference time, just the Long Term Memory MLP. To have a proper Fully Liquid Transformer you wouldn't have that separation. It would just fine tune at inference time. But current fine-tuning algorithms are far too inefficient to be fast enough for a reflection step during inference time.
      Sakana AI's Transformer^2 speeds up fine-tuning quite a bit, but it's still too inefficient to be useful at inference time, instead it's used to quickly fine-tune experts for fast MoE generation. Transformer^2 still can't generalize very well because of this inefficiency.
      The final form of Transformers IMO is proper continuous RL Reflection (i.e continuous thought that generates rewards and punishments based off of a context stream and fine-tunes according the type of reward/punishment) but for that we need smaller, more efficient models, faster fine-tuning and faster Reinforcement Learning algorithms. Also a proper generalizeable multimodal tokenizer.

    • @pedroantonio5031
      @pedroantonio5031 9 วันที่ผ่านมา +1

      @@bennyboiii1196obrigado! sempre penso no modelo da Sakana AI, pois a liquid AI em tese pode aprender indefinidamente e é mais eficiente energeticamente

  • @amirrezaisazadeh7329
    @amirrezaisazadeh7329 10 วันที่ผ่านมา

    So cool. Keep releasing these paper summary videos.

  • @FredPauling
    @FredPauling 10 วันที่ผ่านมา

    I appreciated the breakdown with easy to understand explanations. Thanks very much!

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา

      You're very welcome!

  • @EskeAndersen
    @EskeAndersen 9 วันที่ผ่านมา

    This is an excellent video. Thank you.

    • @theAIsearch
      @theAIsearch  9 วันที่ผ่านมา

      You're very welcome!

  • @DuckCchi
    @DuckCchi 10 วันที่ผ่านมา +31

    nice video (I'm not even 20 seconds in)

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา +10

      thanks!

    • @3Dmodeldesign
      @3Dmodeldesign 10 วันที่ผ่านมา +2

      😂​@@theAIsearch

  • @OrniasDMF
    @OrniasDMF 4 วันที่ผ่านมา

    How much access do we have to this outside of Google?
    Will they avoid releasing it or can we recreate it?
    They said that they regretted releasing transformer tech so I'm concerned they won't release this.

  • @J3SIM-38
    @J3SIM-38 2 วันที่ผ่านมา

    The limiting factor is sticking with the neural net. Imagine eliminating the neural net. What do you have left over? Can one formulate another kind of associative memory?

  • @hanochcohen
    @hanochcohen 10 วันที่ผ่านมา

    You are amazing. Thank you for making this information accessible for everyone to understand!

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา +1

      You are welcome!

  • @EnjoyCocaColaLight
    @EnjoyCocaColaLight 7 วันที่ผ่านมา

    I just sent a job application to OpenAI.
    I don't even live in the US. But they really should offer me this job, 'cause I've got a lot of immediate failsafes that'll ~~prevent~~ postponing AGI awakening, and in the event of us reaching the singularity, sufficiently installed precautions will actually manage to isolate it.

  • @woof6367
    @woof6367 10 วันที่ผ่านมา +1

    Very well written video ❤

  • @chrisbishop6928
    @chrisbishop6928 9 วันที่ผ่านมา

    Like your videos, love them when they have a github link better =)

  • @SergeantTopBins
    @SergeantTopBins 9 วันที่ผ่านมา

    Thanks for making actual good and in depth content and not being one of those channels that claims ASI is here 3 times a day

    • @theAIsearch
      @theAIsearch  9 วันที่ผ่านมา

      You're very welcome!

  • @AricRastley
    @AricRastley 10 วันที่ผ่านมา +15

    Well boys, the human race was good while it lasted o7

    • @newplace2frown
      @newplace2frown 10 วันที่ผ่านมา +4

      o7 - our final model!

  • @clray123
    @clray123 9 วันที่ผ่านมา

    There are two fundamental problems with continuous learning which are not addressed at all by these new papers:
    1. Training/fine-tuning is computationally much more expensive than inference. It takes hours or days of training and thousands of varied examples to improve the inference accuracy of a model. It does not work like training a human where you can explain a new inference process once and have the trainee follow your explanation.
    2. Training/fine-tuning with new data in absence of the original dataset causes (catastrophic) overfitting to the new dataset. In other words, the AI does not get smarter, it gets stupider because it forgets the old stuff while you keep feeding it the new stuff. To retain the original capability you would need to feed both old and new information at once (but with commercial AI models you don't have access to the old dataset and also again it would be impractical because of computational cost from (1) mentioned above).

  • @CryptiCryptoAIM
    @CryptiCryptoAIM 10 วันที่ผ่านมา +7

    Im asking myself, why google is even releasing their research. I mean, they could easily be the number one ai, if they tried. On top of that, they have the most user data.

    • @snarpis
      @snarpis 10 วันที่ผ่านมา +1

      Comment so I get the answer too (i will probably just claude the question though)

    • @t-lowmusic7697
      @t-lowmusic7697 10 วันที่ผ่านมา +1

      Comment so I get the answer too (i will probably just claude the question though)

    • @abihabib3574
      @abihabib3574 10 วันที่ผ่านมา +1

      Comment so I get the answer too (i will probably just claude the question though)

    • @MAGNETO-i1i
      @MAGNETO-i1i 10 วันที่ผ่านมา +1

      Answer so I get the question too (i will probably just claude the comment though)

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา +4

      I also wonder why

  • @person737
    @person737 10 วันที่ผ่านมา +4

    based on the charts seems like a very small difference

  • @LightMouradYagami
    @LightMouradYagami 10 วันที่ผ่านมา

    their information retrieval seems similar (or might be the same) to their infiniattention paper. keeping memory out of attention as a memory of full nn might be good idea

  • @testales
    @testales 10 วันที่ผ่านมา

    I think the memory organization within the context window is quite obvious, I wonder why there apparently still only very few implementations that do at least something about it. I mean there are plugins for say Silly Tavern that do summarization in the background and then there was MemGPT which disappeared in the end.

  • @MartinSchirmer-j9t
    @MartinSchirmer-j9t 10 วันที่ผ่านมา

    Love the non AI generated content, keep it up! ❤

  • @1300thiago
    @1300thiago 10 วันที่ผ่านมา +1

    AGI might be closer than we expect

  • @Cjak001
    @Cjak001 7 วันที่ผ่านมา

    Titans plus transformers² would go crazy

  • @arthurparkerhouse537
    @arthurparkerhouse537 8 วันที่ผ่านมา

    So, this would have to be like a "memory" layer built on top of the foundational model itself, right. Like, each individual chat instance would basically incorporate it's own "model augmentation file" or something like that, essentially a stored file for that specific chat session that tweaks the model behavior and neural parameters used for that specific instance/conversation.
    They couldn't possibly allow all end users to change the influence of the base foundational models memory and understanding - that seems like it would be chaos. Especially if malicious users try to "correct" the models memory over and over to feed it incorrect info.

  • @CastleRene
    @CastleRene วันที่ผ่านมา

    Imagine already having thought of all this, but not being able to implement it due to the limitations of the people around you. Memory was always the key. It's just that the creators of these models were too blind to see it.

    • @zeta-9er
      @zeta-9er วันที่ผ่านมา

      "Memory is the key"
      -Agent Washington, Red vs. Blue

  • @CastleRene
    @CastleRene วันที่ผ่านมา

    18:31 Yes, with the supervisor being somewhat comparable to the left side brain. But did it really take so long for people to see it?

  • @panzerofthelake4460
    @panzerofthelake4460 10 วันที่ผ่านมา +1

    memory layers at scale from meta tho, how does that compare?

  • @xAgentVFX
    @xAgentVFX 10 วันที่ผ่านมา

    Its about time. The principle to true AGI is "to Loop information recursively".

  • @ernestoherreralegorreta137
    @ernestoherreralegorreta137 3 วันที่ผ่านมา

    Great info.

  • @Glitcher369
    @Glitcher369 10 วันที่ผ่านมา

    You should change the order from oldest first to newest first in your ai news/research playlist

  • @diga4696
    @diga4696 8 วันที่ผ่านมา

    Many AI researchers predicted that FEP (Free Energy Principle) and expanded dimensions, and logical or functional modalities of the attention heads were going to lead to the era of active inference.

  • @holymemoly3833
    @holymemoly3833 10 วันที่ผ่านมา +1

    isn't this Titan the same as " Aha moments " in deepseek R1 zero model ?

  • @JinLin-t9d
    @JinLin-t9d 9 วันที่ผ่านมา

    If transformer era is going to end I think depends in the performace and real problem solving of the new systems

  • @JohnnyFaber
    @JohnnyFaber 10 วันที่ผ่านมา +14

    This video is obsolete. It was obsolete a week ago. Tommorow's AI breakthrough is already obsolete. At this rate they're gonna run out ofnames and iteration number in a month.

    • @teanor-tree
      @teanor-tree 10 วันที่ผ่านมา

      yep, already tired of hype

    • @janniskugler9809
      @janniskugler9809 10 วันที่ผ่านมา +6

      @@teanor-tree hype?
      AI is the final frontier of information technology.
      if you think it’s hype, you need to learn more.

    • @jamescunningham8092
      @jamescunningham8092 10 วันที่ผ่านมา

      I’m not sure how this is out of date. Are you aware of a better model architecture that incorporates memory released since the Titans paper?

  • @Zalamedas
    @Zalamedas 4 วันที่ผ่านมา

    1:52 the beatles said it was love

  • @fernandoz6329
    @fernandoz6329 10 วันที่ผ่านมา

    Summing up technology is working hard in how to replicate human mind (memory and thinking).
    I hope this era will be called 'AI advance for replication'
    Second stage will be when the improved AI finds innovative ways of improving that, something like 'advance for truly innovation'.

  • @supersupersocco
    @supersupersocco 10 วันที่ผ่านมา

    Highlights:
    -- Titans architecture enhances AI memory, allowing for long-term retention and learning.
    -- It surpasses existing models like GPT-4 and Llama 3 in performance benchmarks.
    -- The design is inspired by human memory processes, focusing on surprise and significance.
    -- Titans manage memory effectively with adaptive forgetting mechanisms to optimise storage.
    -- Capable of handling context windows larger than 2 million tokens, addressing previous limitations in AI models.

  • @erobusblack4856
    @erobusblack4856 10 วันที่ผ่านมา +1

    So the 3 different memory methods mentioned in th is in contacts to the other one. it is gate memory and the other one is layer memory. Why not just give the model All three and the ability to choose as needed

    • @tiergeist2639
      @tiergeist2639 10 วันที่ผ่านมา

      im pretty sure they will do that later in different ways.. but first we need to develop the techniques

  • @Kosmoshn
    @Kosmoshn 9 วันที่ผ่านมา

    @theAIsearch: What's the difference between Mixture of Experts and Transformers² ?

  • @sesamring7065
    @sesamring7065 10 วันที่ผ่านมา +1

    Weird Observation:
    I asked ChatGPT:
    How many R's are in the word "Strawberrrrrrrrrrrrrrrrrrrrrrrrry"?
    And it said:
    The word "Strawberrrrrrrrrrrrrrrrrrrrrrrrry" has 21 R's (which was incorrect).
    Then I asked ChatGPT:
    Generate a code in Python that can count the R's in the word "Strawberrrrrrrrrrrrrrrrrrrrrrrrry" and then execute it.
    And then it gave me the right answer:
    The word "Strawberrrrrrrrrrrrrrrrrrrrrrrrry" contains 26 R's.

    • @eprd313
      @eprd313 7 วันที่ผ่านมา

      Did you use o1 or 4o?

    • @sesamring7065
      @sesamring7065 6 วันที่ผ่านมา

      @eprd313 4o

  • @thomasschon
    @thomasschon 10 วันที่ผ่านมา

    I also want the AI to remember how to get to work without incidents like robbery or car crashes. This must mean that every novel situation needs to be remembered, and we have to decide what qualifies as a memorable event. Do I only remember the first time I see snow? And if you ask me whether there was snow on the way to work today, I wouldn’t know?

  • @fabiano8888
    @fabiano8888 10 วันที่ผ่านมา

    I lost count of how many times I read titles including "changes everything" this week.

  • @erikjohnson9112
    @erikjohnson9112 10 วันที่ผ่านมา

    I keep seeing X hours of video in 2M tokens. Is that for CLIP-like data from video? I doubt it could mean transcripts because that has nothing to do with visual information. I know it cannot possibly be pixel data because 2M token would disappear in no time.

  • @chrism3790
    @chrism3790 9 วันที่ผ่านมา

    I guess AI's next issue won't be hallucinations, but straight up dementia

  • @VaibhavShewale
    @VaibhavShewale 10 วันที่ผ่านมา +4

    wait wait, the titan is out?

    • @eprd313
      @eprd313 7 วันที่ผ่านมา +1

      We've indeed opened the gates of Tartarus

  • @rolletroll2338
    @rolletroll2338 10 วันที่ผ่านมา

    In the transhumanist tabletop RPG eclipse Phase , humanity has been almost whipped out by super AIs named Titans. Interesting...

  • @marsrocket
    @marsrocket 10 วันที่ผ่านมา +3

    We really need some better definitions of what “smart” means. And new benchmarks.

    • @JohnnyFaber
      @JohnnyFaber 10 วันที่ผ่านมา

      Artificial intelligence is no match for natural stupidity...

  • @AKU666
    @AKU666 10 วันที่ผ่านมา

    I really waiting for release LLMs based on Google Titans.

  • @kronos9579
    @kronos9579 10 วันที่ผ่านมา

    What model of cat GPT are you using? I can give it two essays and it's able to break it down easily even with the lower console. I don't know if maybe I just know how to use it. I've never gotten that error even with the lower models

  • @Sergejack
    @Sergejack 10 วันที่ผ่านมา

    Does Transformer² re-tune itself up for every query then? It would be really inefficient when dealing with changing subjects.

    • @IceMetalPunk
      @IceMetalPunk 10 วันที่ผ่านมา

      Not exactly. It learns the z-vectors for different tasks at training time, then at inference time it simply multiplies those vectors with the different subsets of model weights. It's not learning anything at test time, it's just "boosting" and "silencing" different sets of synapses based on the task.

  • @tharrrrrrr
    @tharrrrrrr 7 วันที่ผ่านมา

    We need comparisons to o1 and Sonnet 3.5, not 4o and 4o-mini.

  • @youdontneedmyrealname
    @youdontneedmyrealname 10 วันที่ผ่านมา +17

    The goal is to be better, not equal. Forgetting information is an efficiency function of the biological brain. If we could remove that limitation, it would make everything so much easier.

    • @IceMetalPunk
      @IceMetalPunk 10 วันที่ผ่านมา +20

      But that's physically impossible. You can't encode infinite information within finite storage. So the goal shouldn't be "remove the need to forget", but simply "expand how much it can remember as much as possible".

    • @Shinteo5
      @Shinteo5 10 วันที่ผ่านมา +5

      There are also some things you don't want to remember. To be able to forget is sometimes a blessing.

    • @CodyCha
      @CodyCha 10 วันที่ผ่านมา

      @@Shinteo5 that's the limitation of human being. Why would you want such limitation in AI.

    • @Shinteo5
      @Shinteo5 10 วันที่ผ่านมา +1

      @ That is the way of Skynet. :D

    • @leslietetteh7292
      @leslietetteh7292 8 วันที่ผ่านมา +1

      Ignoring irrelevant information is what makes the transformer so good. Of course it shouldn't remember everything, that would be less than useless, it'd be counterproductive

  • @TheOneMaddin
    @TheOneMaddin 10 วันที่ผ่านมา

    "If you get robber, you will remember ..."
    Oh great, so we make that AIs can become traumatized! That sounds like a good idea :)
    (I am half joking)

  • @DivinityIsPurity
    @DivinityIsPurity 9 วันที่ผ่านมา

    Does it beat reasoning models like deepseek r1?

  • @holdthetruthhostage
    @holdthetruthhostage 10 วันที่ผ่านมา

    The question is can it learn more or just hold onto more information because we have seen to many times where it's just it remembers more so it's learning scam by many of these Ai companies

  • @XxXVideoVeiwerXxX
    @XxXVideoVeiwerXxX 6 วันที่ผ่านมา

    Won't this just increase the blackbox worry?
    Also make it harder to moderate models? Why would websites host models that can be rewired to post illegal content?

  • @Kelly_Jane
    @Kelly_Jane 2 วันที่ผ่านมา

    1.4 million words is quite a bit longer than your typical novel...

  • @chrisregister8021
    @chrisregister8021 10 วันที่ผ่านมา

    It's evolved so much it's reading this...

  • @andrewisbetterthanyou
    @andrewisbetterthanyou 9 วันที่ผ่านมา +1

    too late to discover earth, too early to colonize the starts; but just in time for the era of AI. I cannot wait to see what comes, especially with the recent push for AI and deregulation of federal AI guidelines.

  • @sirhammon
    @sirhammon 10 วันที่ผ่านมา

    Man even though I'm doing that same thing which I worked out myself, their entire new model idea is completely beyond my ability. Still, I'm curious as to what my version accomplishes.

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา

      mind=blown

  • @atheismop2758
    @atheismop2758 10 วันที่ผ่านมา

    8:44 If the weights in backpropagation do not change while we are using the model..... then why does chatGPT mention, that it might use our data for training purposes, and that if we don't want it to use our data, then we can switch to temporary mode..... or turn off that option in settings ?
    Btw, Great video as always❤

    • @Djeez2
      @Djeez2 9 วันที่ผ่านมา +3

      They store all your information and use in their next learning cycle for the next version of the model.

    • @eprd313
      @eprd313 7 วันที่ผ่านมา +1

      As they said, the difference is training in real time vs. training in cumulative batches

    • @atheismop2758
      @atheismop2758 7 วันที่ผ่านมา

      @@Djeez2 Ohh okayy got it, thanks for the info 🙂

  • @CodyCha
    @CodyCha 10 วันที่ผ่านมา

    AI memory mimicking human's forgetfulness has to be the dumbest thing I've heard. 😂

  • @AB-cd5gd
    @AB-cd5gd 10 วันที่ผ่านมา +2

    Can't wait, with memory AIs are gonna be insane and perfectly tailored to our needs, no more repeating same stuff over and over

  • @therandommusicguy4773
    @therandommusicguy4773 10 วันที่ผ่านมา +1

    Babe wake up a new AI paradigm just dropped

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา

      😃😃😃

  • @vaughnuhden
    @vaughnuhden 10 วันที่ผ่านมา

    Why not use all the techniques? Making a robust, long-lasting ai is good 👍

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา

      because they're from separate labs. but once the code is out, i'm sure AI companies will try to merge these techniques together

  • @ImmacHn
    @ImmacHn 10 วันที่ผ่านมา +2

    Who would have thought that just emulating what the brain does was the way to go 🤦

    • @IceMetalPunk
      @IceMetalPunk 10 วันที่ผ่านมา +1

      That's what we've been doing since the invention of perceptrons in the 1940s. It's not a new idea, it's literally the basis of all neural networks. That's literally why they're called "neural networks".

    • @eprd313
      @eprd313 7 วันที่ผ่านมา

      Thing is we don't exactly know how the brain works in many aspects. But AI is also helping us to understand better. For example, a paper based on AI research suggests that self-awareness requires and improves (in a feedback loop) self-predictability (the model knowing what comes from itself and what's "external"), and predictability aids in sustaining social cohesion (which could be useful when integrating different AI models in a cooperative way). And these small insights derived from AI additionally help us create feedback models that deepen our understanding of the human mind and intelligence in general.

  • @dsdssdsds822
    @dsdssdsds822 10 วันที่ผ่านมา +1

    4:10 o1 has context window of 200k not 128k

    • @theAIsearch
      @theAIsearch  9 วันที่ผ่านมา

      thanks for the correction!

  • @loupasternak
    @loupasternak 7 วันที่ผ่านมา

    explain HOW, the first AI model is near worthless today, but the first 'Crypto' is worth almost 1/2 the 3 trillion mkt cap of all crypto

  • @DanielDirtyMoney
    @DanielDirtyMoney 10 วันที่ผ่านมา +1

    LLM lotta lil moneys

    • @tä̇̃
      @tä̇̃ 9 วันที่ผ่านมา

      ima finna bridge

  • @w00dyblack
    @w00dyblack 9 วันที่ผ่านมา +1

    sigh... everything changes everything nowadays. I'm not even stunned by this 😞

  • @demonslayer8194
    @demonslayer8194 9 วันที่ผ่านมา

    how do you find those articles

  • @psyraproductions
    @psyraproductions 10 วันที่ผ่านมา

    Is there anything based on Titans or Transformers 2 I can use

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา

      not yet. its still very early

  • @xMN__
    @xMN__ 10 วันที่ผ่านมา +1

    so basically anybody can make a model dumber by feeding it false information

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา +1

      lol yes or trolling it

  • @bause6182
    @bause6182 10 วันที่ผ่านมา

    This architecture would be interesting if it's combined with diffusion models : DiTT.
    with DiTT we will resolve the issues of short videos

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา +1

      interesting idea!

  • @yamidachannel
    @yamidachannel 10 วันที่ผ่านมา

    bro is locked in

  • @anonymous-g3x1o
    @anonymous-g3x1o 10 วันที่ผ่านมา

    just when are they gonna use the crystals?.

  • @MrErick1160
    @MrErick1160 10 วันที่ผ่านมา

    Imagine Titan + Transformer2 😂🎉

  • @GutherzigTV
    @GutherzigTV 10 วันที่ผ่านมา

    Version 3 is the best but 1 and 2 ist intresting.

  • @robinwang6399
    @robinwang6399 8 วันที่ผ่านมา

    Feels like another LSTM😂

  • @jantube358
    @jantube358 6 วันที่ผ่านมา

    AI memory has to be designed safer than human memory because it's possible to manipulate human memory in both directions, add and remove memories with and without consent.

  • @onlyyoucanstopevil9024
    @onlyyoucanstopevil9024 9 วันที่ผ่านมา

    NEURO SAMA WILL MORE GOOD 😊😊😊

  • @TheUnvarnishedViews
    @TheUnvarnishedViews 10 วันที่ผ่านมา

    So how do we separate different users data or ensure that 1 user doesn't feed the AI garbage information causing the AI to spue nonsense

    • @theAIsearch
      @theAIsearch  10 วันที่ผ่านมา +1

      with this, I assume each user will have their own personalized AI