Transformer LLMs are Turing Complete after all !?

แชร์
ฝัง
  • เผยแพร่เมื่อ 24 พ.ย. 2024

ความคิดเห็น • 32

  • @DerPylz
    @DerPylz 2 หลายเดือนก่อน +16

    This was super nice! I hope AI Coffee Break podcasts become a regular thing!

    • @AICoffeeBreak
      @AICoffeeBreak  2 หลายเดือนก่อน +3

      AI Coffee Break pods.
      Thank you! It was harder to do a podcast than expected.

    • @DerPylz
      @DerPylz 2 หลายเดือนก่อน +3

      @@AICoffeeBreak

  • @marzi869
    @marzi869 2 หลายเดือนก่อน +5

    I should read the paper first, then watch this amazing talk again. I did not understand completely. But thank you for this precious content.

  • @hiauoe
    @hiauoe 2 หลายเดือนก่อน +10

    Wonderful podcast, perfect duration, great questions, keep them coming!

  • @shadysaeed644
    @shadysaeed644 2 หลายเดือนก่อน +2

    Great to see you, I remember you from that beautiful video you made couple of years back on the "attention is all you need" paper

  • @CodexPermutatio
    @CodexPermutatio 2 หลายเดือนก่อน +4

    Great content and format. This channel just keeps getting better.

  • @kevon217
    @kevon217 2 หลายเดือนก่อน +3

    Fascinating. Great discussion.

  • @gergelyszabo4802
    @gergelyszabo4802 2 หลายเดือนก่อน +7

    Being Turing complete is nice and all, but this does not inherently mean that there will be existing solutions to certain problems inside reasonable computational limitations.

  • @GNARGNARHEAD
    @GNARGNARHEAD 2 หลายเดือนก่อน +3

    great video thanks 😁

  • @ZafyArimbola
    @ZafyArimbola 2 หลายเดือนก่อน +1

    Can His tutorial on computational expressivity of Language Model be found on youtube?

    • @AICoffeeBreak
      @AICoffeeBreak  2 หลายเดือนก่อน +1

      I'm afraid not. Only written form and slides here: acl2024.ivia.ch/
      acl2024.ivia.ch/about

  • @thomasmuller7001
    @thomasmuller7001 2 หลายเดือนก่อน

    its funny, in the universal transfromer paper they show turing completeness by ignoring the decoder while focusing on the adaptive compute time layerwise instead of context-step-wise "chain of thought" .. so transformers are double-turing-complete? :D

  • @rockapedra1130
    @rockapedra1130 2 หลายเดือนก่อน

    21:08 Great! Maybe we need an AI so smart that we can ask "solve the problem but make sure you do it in a way that our puny human brains can understand". Weird world we're creating.

  • @psi4j
    @psi4j 2 หลายเดือนก่อน

    🎉

  • @mrstephanwehner
    @mrstephanwehner 2 หลายเดือนก่อน +1

    Contradicts your last video about Chomsky's assertion? th-cam.com/video/8lU6dGqR26s/w-d-xo.htmlsi=K26pFdgNo35-_-hu Meaning Chomsky' was correct, no? Your question at 18:00 goes into this a little

  • @andresot
    @andresot 2 หลายเดือนก่อน

    Didn't watch yet - at work xD. Did they solve the infinite input width problem?

    • @AICoffeeBreak
      @AICoffeeBreak  2 หลายเดือนก่อน +6

      Did Turing solve the problem with the infinite tape for the Turing machine?
      We talk in the video about whether anything in reality (e.g., my laptop) is equivalent to a turing machine. 😅

    • @andresot
      @andresot 2 หลายเดือนก่อน +1

      @@AICoffeeBreak Why should math care about my laptops memory? That argument is nonsensical.

    • @AICoffeeBreak
      @AICoffeeBreak  2 หลายเดือนก่อน +4

      Because Turing machines are just a mathematical theory requiring infinite resources. Infinity does not exist for practical reasons in the universe and we can use only finite resources. So in the same way in which laptops would need infinite memory to implement Turing Machines, neutral networks would need infinitely wide vectors or infinite precision.

  • @haniamritdas4725
    @haniamritdas4725 2 หลายเดือนก่อน +4

    A ticket tape machine is Turing Complete. It's a very low bar

    • @AICoffeeBreak
      @AICoffeeBreak  2 หลายเดือนก่อน +6

      That Transformer Encoders do not even meet. :)

  • @genkidama7385
    @genkidama7385 2 หลายเดือนก่อน

    llms can generate coherent sentences, they are probabilistic in nature, and based on randomness. randomness/random sampling of the latent space creates halucinations, because there is no reflection, no cognition of any kind hapening. a very advanced automaton, that can imitate humlan language, nothing else, pure imitation. just like a plane isnt a bird, it flies, it goes superfast, but its not a bird. similarly a language model imitates words and sentences, but there is nothing else going on.

    • @jcoffland
      @jcoffland 2 หลายเดือนก่อน

      It is true that LLMs use randomness and statistical mathematics but that does not necessarily mean they are limited to only producing random statistical samplings. The whole is not always just the sum of the parts. LLMs can succeed in some limited reasoning. For example, it has been shown that simply asking an LLM to "think step-by-step" causes it to produce objectively better answers. How could this work if the LLM were not reasoning? If it was simply reproducing patterns you might as well just ask it to "give me a smarter sounding answer." LLMs are also able to solve logic and mathematical problems. Successfully finding solutions to such problems will not happen consistently by random chance or by simply producing textual patterns that look like answers. Some level of reasoning is required.
      The hallucination problem occurs when you ask an LLM to produce an answer outside of its capabilities. Instead of saying, "I don't know," it will often make up something that looks like an answer. But humans do this all the time too and that does not mean we are incapable of reasoning.
      It's easy and reassuring to dismiss LLMs as statistical parrots because they are then clearly not a threat. I don't believe LLMs alone, at least in their current form, pose a significant direct threat to humans. The only real danger is in how humans use LLMs for good or bad purposes. We are still a long way from AGI but these are exciting times.

    • @zwanzikahatzel9296
      @zwanzikahatzel9296 2 หลายเดือนก่อน

      @@jcoffland I think ppl conflate various types of reasoning. there is a type of reasoning that is purely mechanical, true and true = true, false or true = true, true xor true = false etc. you can memorize and do a lot of "reasoning" just by applying rules or patterns mechanically, you don't even need to know why each rule is true. classical computers and simple logic gates can already carry out this type of reasoning, thats why we've had theorem provers and prolog for decades. LLMs can do something similar but in a more statistical/probabilistic/fuzzy way. they can combine truthfulness of statements and draw conclusions probabilistically, so it looks like it has some common sense reasoning, but it;s not, it's just expanding the type of logic gate reasoning to linguistic representations. The other type of reasoning people talk about is about how the actual world works and cause/effect and intentionality. For example when I say "the cat is on the table" you dont just "compute" sentence linguistically/symbolically, but also visually, you imagine a cat on a table, because of how physics works you know the can cannot also be underneath at the same time, u know the cat could jump off at any point. you are not just thinking about that scene linguistically, like an LLM that has never actually seen a cat. its as expecting a human who has never seen, touched or heard anything and only been able to interact with a screen of printed words and a keyboard to be able to understand the world. maybe this will change once ai is able to process data in real time, is embedded in a robotic body and interact with the world, until then llms are not actually able to understand anything, they only have a purely linguistic intelligence/understanding (they manipulate words and symbols statistically)

  • @andriik6788
    @andriik6788 2 หลายเดือนก่อน +3

    Frodo! Is that you?!

    • @CharlesNiko-j3p
      @CharlesNiko-j3p 2 หลายเดือนก่อน

      Most likely, the phoenix from Gladiator

    • @doomsday_netrunner
      @doomsday_netrunner 2 หลายเดือนก่อน

      no that's eddie from limitless

  • @Paplu-i5t
    @Paplu-i5t 2 หลายเดือนก่อน +2

    Lip stick, east european accent, and AI.

  • @TheHouseOfBards
    @TheHouseOfBards 2 หลายเดือนก่อน

    They are beyond completeness as their attention mechanism esentially gives them a self destruction switch that flips the context and frame of a problem as it relative hardness requires it.
    New Oracle Turing Machines could arise from the attention mechanism. Transformers should definitely be studied further by CS theorists.