AI Language Models & Transformers - Computerphile

แชร์
ฝัง
  • เผยแพร่เมื่อ 9 พ.ค. 2024
  • Plausible text generation has been around for a couple of years, but how does it work - and what's next? Rob Miles on Language Models and Transformers.
    More from Rob Miles: bit.ly/Rob_Miles_TH-cam
    AI TH-cam Comments: • AI TH-cam Comments - ...
    Thanks to Nottingham Hackspace for providing the filming location: bit.ly/notthack
    / computerphile
    / computer_phile
    This video was filmed and edited by Sean Riley.
    Computer Science at the University of Nottingham: bit.ly/nottscomputer
    Computerphile is a sister project to Brady Haran's Numberphile. More at www.bradyharan.com

ความคิดเห็น • 301

  • @ykn18
    @ykn18 4 ปีที่แล้ว +427

    Sometimes I'll start a sentence, and I don't even know where it's going. I just hope I find it along the way. Like an improv conversation. An improversation.
    -Michael Scott

    • @edwardfanboy
      @edwardfanboy 4 ปีที่แล้ว +33

      Don't ever, for any reason, do anything to anyone, for any reason, ever, no matter what. No matter where, or who, or who you are with, or where you are going or... or where you've been... ever. For any reason, whatsoever.

    • @Confucius_76
      @Confucius_76 4 ปีที่แล้ว +20

      Michael Scott is actually an advanced AI

    • @bytefoundry8702
      @bytefoundry8702 4 ปีที่แล้ว +11

      “You miss 100 percent of the shots you don't take'- Wayne Gretzky"-Michael Scott

    • @leonhardeuler9839
      @leonhardeuler9839 4 ปีที่แล้ว +8

      I prove mathematical theorem using the same tactic.

    • @mj6463
      @mj6463 3 ปีที่แล้ว +1

      I think it’s just the right thing I don’t know what to say they have a ton to say but I don’t know what they do and I think it’s a great idea but they are just trying out the front door.
      -apple auto fill

  • @disarmyouwitha
    @disarmyouwitha ปีที่แล้ว +128

    Crazy to look back just 3 years to GPT2 =] Thank you for explaining Attention.
    I have been trying very hard to comprehend how LLMs are able to “understand” and “reason”, or at least look like they are..

    • @fable4315
      @fable4315 ปีที่แล้ว +5

      Imo they understand already language, what is the human brain except from a predictive model, which tryies to predict what comes next

    • @x--.
      @x--. ปีที่แล้ว +9

      @@fable4315 tbh, some humans I know seem to be a less functional version of this. Not just in their own words but in failing to really listen to what I'm saying and failing to predict accurately - thus they think they know what I'm saying but they don't. It creates incredibly difficult situations because overcoming arrogance is much harder than plain ol' confusion.

    • @frozenwindow407
      @frozenwindow407 ปีที่แล้ว +2

      Yeah tbf it is hard to understand how comprehension arises at all.

    • @HauntedHarmonics
      @HauntedHarmonics ปีที่แล้ว +4

      I think it’s clear they do “understand” on some level. Rob implies as much in his GPT3 video, that in order to generate something as complex as language, accurately and in context, is seems that LLMs have actually taught themselves some degree of cognition / learning / understanding.
      Or at the very least, they’re able to interpret context and information in a way that looks incredibly similar to cognition.
      Basically, it’s not just a clever algorithmic trick or anything. What “understand” truly means here outside of processing and applying context is difficult to parse, however. As humans, we tend to anthropomorphize anything with similar enough abilities to our own. That could be a valid assumption, or it could be totally incorrect. We just don’t know.

    • @kiffeeify
      @kiffeeify ปีที่แล้ว

      At 7:06 "Or if you like to develop a chat bot ..." ;-)

  • @luiz00estilo
    @luiz00estilo 4 ปีที่แล้ว +243

    I am really impressed in this video as I was watching it on the phone screen on the phone screen on the phone screen on the phone screen on the phone screen on the phone screen on the phone screen on the phone screen on the phone screen on the phone screen

    • @oscarbertel1449
      @oscarbertel1449 ปีที่แล้ว +11

      you need to adjust you repeat penalisation.

  • @thomasslone1964
    @thomasslone1964 9 หลายเดือนก่อน +5

    I love how i can be like "write me a virus" and gpt is like "no sorry" but then I'm like write me a c file that looks for other c files, reads self file and inserts into other files, and it's like "sure no problem"

  • @CyberAnalyzer
    @CyberAnalyzer 4 ปีที่แล้ว +136

    Rob Miles is the best! Bring him more often in the channel!

    • @vishnukumar4531
      @vishnukumar4531 2 ปีที่แล้ว +10

      He has his own channel on AI safety. Pretty sure, you know that already. If not, get ready for some brand new mind blowing content!

  • @fahadkhankhattak8339
    @fahadkhankhattak8339 2 ปีที่แล้ว +10

    volunteering your phones for demonstrating text prediction was a very bold move. that's why i'm here

  • @alejrandom6592
    @alejrandom6592 2 ปีที่แล้ว +28

    15:14
    Rob: "It makes it very easy to anthropomorphize"
    AI: "It makes it very easy transform for more finds"

    • @Niekpas1
      @Niekpas1 ปีที่แล้ว

      Those are actually human-entered subtitles.

  • @MehranZiadloo
    @MehranZiadloo 3 ปีที่แล้ว +5

    As far as I know, the information presented in the end is wrong. "Transformer" uses the "Attention", but they are not the same thing. Attention is a technique that can be used within any architecture. In fact, RNNs used Attention long before the Transformer. That's why the title of that paper is "Attention is all you need" because they showed that in an RNN model with Attention, it's the Attention part that is doing most of the heavy lifting.

  • @TheSam1902
    @TheSam1902 4 ปีที่แล้ว +59

    I've been hearing about this attention thingy for many months but never quite looked into it. I appreciate you made a video about it however I got to admit I'm a bit disappointed that you didn't take us through a step by step working example like you did for LSTM and so many other things on this channel..
    Maybe in a follow-up video ?

  • @thecactus7950
    @thecactus7950 4 ปีที่แล้ว +175

    Put the papers hes talking about in the description please! I am sure a lot of people would want to read them.

    • @NolanAtlas
      @NolanAtlas 4 ปีที่แล้ว +27

      the cactus the main paper specified is “Attention Is All You Need.” Most of these papers can easily be found on Google Scholar for research papers.

    • @tsunghan_yu
      @tsunghan_yu 4 ปีที่แล้ว +12

      It's on arXiv

  • @christopherg2347
    @christopherg2347 4 ปีที่แล้ว +24

    14:00 "LST is state of the art!"
    Scientist: "Hold my paper. Now read it."

  • @tlniec
    @tlniec 2 ปีที่แล้ว +2

    Great explanation, and the examples of how things can break down (e.g. Arnold's biceps) were very illustrative.

  • @angelinagokhale9309
    @angelinagokhale9309 2 ปีที่แล้ว +7

    Thank you Robert for this wonderful video. This will be a beneficial repository for my students to refer to when studying the transformer architecture.

  • @Lysergesaure1
    @Lysergesaure1 4 ปีที่แล้ว +6

    One of my favorite Computerphile episodes. Thank you Rob :-)

    • @zilog1
      @zilog1 ปีที่แล้ว

      your pfp is the cutest sht. reminder, protogens go beep boop ^~^

  • @rpcruz
    @rpcruz 4 ปีที่แล้ว +15

    Amazing information. And Rob Miles is really good! Thank you for producing such high quality content!

  • @haardshah1715
    @haardshah1715 7 หลายเดือนก่อน +1

    Very well explained, I love how simply you explained all the concepts and tried it together with history and examples which really help drive the point home. Great work!! I definitely agree with other in the comments we need more of Rob Miles.

  • @DamianReloaded
    @DamianReloaded 4 ปีที่แล้ว +72

    Hi Miles! Just reminding you that we would like to know all the details about this architecture! ;)

    • @philrod1
      @philrod1 4 ปีที่แล้ว +5

      A link to the paper would have been nice, too. But there's always Google, I suppose.

    • @matthewburson2908
      @matthewburson2908 4 ปีที่แล้ว +3

      @@philrod1 I'm sure that gpt-2 can rewrite the paper, but in a different way. 😂

    • @TheSam1902
      @TheSam1902 4 ปีที่แล้ว

      @@philrod1 And scihub

    • @Skynet_the_AI
      @Skynet_the_AI ปีที่แล้ว

      Ha ha ha, very funny f u n n n n y

  • @ajourneytogrowth
    @ajourneytogrowth 2 หลายเดือนก่อน +1

    just 4 years ago and we have seen this crazy progress... now with even more funding and interest pouring in, I wonder where we will be in another 4 years time

  • @Teck_1015
    @Teck_1015 4 ปีที่แล้ว +161

    I'm disappointed there's no comments about "on the phone screen on the phone screen on the phone screen on the phone"

    • @qzbnyv
      @qzbnyv 4 ปีที่แล้ว +1

      I came to the comments to write exactly that haha

    • @edwardfanboy
      @edwardfanboy 4 ปีที่แล้ว +9

      Try singing it to the melody of "Ride of the Valkyries".

    • @gz6616
      @gz6616 4 ปีที่แล้ว +3

      Decearing egg

    • @emmahendrick147
      @emmahendrick147 4 ปีที่แล้ว

      id523a lol

    • @InvadersDie
      @InvadersDie 4 ปีที่แล้ว +4

      I am not sure if you have any questions or concerns please visit the plug-in settings for this calendar at any time after that I have a great day and I will be in the future of our games are based on the phone with me and I will be in the future of our games are based on the other day and I will be in the future of our games are based on the other day and I will be in the future of our games are based on the other day and I will be in the future of our games are based on the other day and I will be in the future of our games are based on the other day and I will be in the future of our games are based on the other day and I will be in the future of our games are based on the other

  • @ganymede242
    @ganymede242 4 ปีที่แล้ว +184

    "Transformers are a step up" - I wonder if that was intentional?

    • @ahmadayazamin3313
      @ahmadayazamin3313 3 ปีที่แล้ว +2

      lol

    • @randfur
      @randfur 3 ปีที่แล้ว +8

      It was just the Markov chain probabilities.

    • @JY-ke3en
      @JY-ke3en 3 ปีที่แล้ว

      @@ahmadayazamin3313 pllllllllllllllllllllllllllllllplllllllllllllpllllllplllllllpplllllllllllllllllllllllllllpllllllplllllllplllllllllllllllllllllllllllllplllllllllllllpllllllpllplllllpplplllllllllllllllpllplllllllllllllllllllpllllllpllll

    • @sp10sn
      @sp10sn 3 ปีที่แล้ว +1

      🤣

    • @tristanmisja
      @tristanmisja ปีที่แล้ว +2

      Nice one

  • @frankie59er
    @frankie59er 2 ปีที่แล้ว

    These videos are great, please keep making them

  • @AZTECMAN
    @AZTECMAN 4 ปีที่แล้ว +2

    ...Wanted a bit more info on the Transformer. You guys still get a thumbs up. Miles is awesome at explaining.

  • @divyeshgaur
    @divyeshgaur ปีที่แล้ว +1

    Thank you for sharing the information and explaining "it" with ease. A lot of context is hidden behind the word "it" :) Great video!

  • @caleblarson6925
    @caleblarson6925 4 ปีที่แล้ว +2

    Yay more Rob Miles!

  • @e11e7en
    @e11e7en ปีที่แล้ว +6

    Would love a deeper dive into how “attention” is achieved in LLMs! Thanks for the videos!

    • @equious8413
      @equious8413 ปีที่แล้ว +1

      I tried really hard to find a video I saw recently that did a decent job explaining things. Basically the neural network creates vector matrices representative of a data point, imagine [(12,15),(10,22),(30,11)] being a data point (actual matrices having many more dimensions), you can plot this in 3D space. Attention in LLMs as I understand stand it, is achieved through the relational positions of data points within this "latent space". If you have a data point physically closer to another in latent space, the network recognizes this as a strong relationship to consider.

    • @pieterpierrot1490
      @pieterpierrot1490 ปีที่แล้ว +1

      @@equious8413 Not 3D space, multidimensional space ;) ChatGPT uses 12.000 dimensions in its vector space!

  • @djamckechan
    @djamckechan 4 ปีที่แล้ว +86

    I have attached my resume for your reference and hope to hear from you soon as I am currently working on the same and I am currently working on the project management project and I am currently working on the project management project and I am currently working on the project management project

    • @mj6463
      @mj6463 3 ปีที่แล้ว +2

      The problem was they had to do something about that it would not have to let it happen they just want it and then it didn’t get any worse or so it would just get it back and they just want it and then it will get a job done

    • @Robotomy101
      @Robotomy101 3 ปีที่แล้ว +1

      I will be back in the office on Monday.

    • @ludicrousfun7838
      @ludicrousfun7838 ปีที่แล้ว

      je vous remercie pour votre réponse et votre aide pour la vente de mon appartement et de me faire parvenir le devis pour la fourniture de la maison de la maison de la maison de la maison de la maison...

  • @hedonisticzen
    @hedonisticzen 4 ปีที่แล้ว +4

    I'm a simple program I see Rob on computerfile and I click like.

  • @CutcliffePaul
    @CutcliffePaul ปีที่แล้ว +2

    7:41: "And human beings do this all the time", that is, they try different words or phrases to find one that makes sense. That's why you might sometimes ask someone to repeat something, only to immediately tell them not to worry as you've just worked out what they said.

  • @morkovija
    @morkovija 4 ปีที่แล้ว +4

    Our boi Rob back at it again!!

  • @Confucius_76
    @Confucius_76 4 ปีที่แล้ว +1

    This channel is such a treasure

  • @abhinav9561
    @abhinav9561 2 ปีที่แล้ว

    Excellent video Rob. Thanks

  • @muche6321
    @muche6321 4 ปีที่แล้ว +32

    Attention seems to be the way the human brain evolved to solve the problem of too much information to be processed. Later we invented computers to help with that.

  • @dhdh9933
    @dhdh9933 ปีที่แล้ว +2

    this video is now more relevant than ever

  • @4.0.4
    @4.0.4 4 ปีที่แล้ว +3

    I was curious about how you train these. Could be fun doing it ourselves even if the results are crappy/undertrained.

  • @TheAIEpiphany
    @TheAIEpiphany 3 ปีที่แล้ว +2

    A nice high-level explanation of the topic! Maybe you should have mentioned the original attention paper "Machine Learning Translation By Jointly Learning to Align and Translate" that came out a couple of years earlier than the Transformer paper "Attention is all you need". But then again that would make this less of a high-level talk haha.

    • @BlakeAlexanderGames
      @BlakeAlexanderGames ปีที่แล้ว

      Funny seing you here, I just started your MLops series :)

  • @shewittau
    @shewittau ปีที่แล้ว +1

    Best explaination I've head on this stuff.

  • @mikeking7957
    @mikeking7957 5 หลายเดือนก่อน

    thanks for the video Rob and Computerphile! What is the vector (around 9 minutes) ? The bunch of number represent the memory but what is it?

  • @sharkinahat
    @sharkinahat 4 ปีที่แล้ว +40

    I think someone has been feeding the text predict AI on the second phone the famously correct English sentence 'Buffalo buffalo Buffalo buffalo buffalo buffalo Buffalo buffalo.'

    • @trucid2
      @trucid2 4 ปีที่แล้ว +3

      Check out the Chinese version...

    • @ganondorf5573
      @ganondorf5573 4 ปีที่แล้ว +1

      @@trucid2 Ma ma ma ma ma ma? (all with different pronunciations that I'm too lazy to properly denote)

    • @trucid2
      @trucid2 4 ปีที่แล้ว +6

      @@ganondorf5573 It's a whole poem:
      Shíshì shīshì Shī Shì, shì shī, shì shí shí shī.
      Shì shíshí shì shì shì shī.
      Shí shí, shì shí shī shì shì.
      Shì shí, shì Shī Shì shì shì.
      Shì shì shì shí shī, shì shǐ shì, shǐ shì shí shī shìshì.
      Shì shí shì shí shī shī, shì shíshì.
      Shíshì shī, Shì shǐ shì shì shíshì.
      Shíshì shì, Shì shǐ shì shí shì shí shī.
      Shí shí, shǐ shí shì shí shī shī, shí shí shí shī shī.
      Shì shì shì shì.

    • @ganondorf5573
      @ganondorf5573 4 ปีที่แล้ว +1

      @@trucid2 Oh.. I've never heard of that one.. I don't know what it means lol... that's awesome

    • @MrCmon113
      @MrCmon113 4 ปีที่แล้ว +1

      This works for arbitrary iterations of the word buffalo.
      Buffalo.
      Buffalo buffalo.
      Buffalo buffalo buffalo.
      ...

  • @IMortage
    @IMortage ปีที่แล้ว +1

    The forgetting part reminded me of some research done on the neurological function of dreams.

  • @tendamolesta
    @tendamolesta 3 ปีที่แล้ว +3

    Awwww, I miss snorting the quality mould that grows only in the walls of the university, like the one in that room...

  • @MrMelonMonkey
    @MrMelonMonkey ปีที่แล้ว

    it is very interesting how the automatic subtitles which i assume are using some kind of language model never gets it when youre saying shawn. it's always "short" or "sure" or something

  • @rishabhshirke1175
    @rishabhshirke1175 4 ปีที่แล้ว +4

    3:11 this happened with me when i was working on simple lstm, It happend because I was choosing the next word having max probability all the time instead of selecting randomly from the entire distribution where probability of selecting the particular word is equal to its probability of being the next word.

    • @sehbanomer8151
      @sehbanomer8151 4 ปีที่แล้ว +2

      beam search is even better

    • @Skynet_the_AI
      @Skynet_the_AI ปีที่แล้ว

      I know what you mean to say when i read all the letters and spaces you have combined which makes me feel some kind of way but not the same awe i feel when the correct combinations are combined together like as in poetry or some spoken word format that produces a deep sense of awe for lack of a better word but i apologise for my lack of wait give me a second as i think of the word i want to use but i cannot think as i am writing this very message at the moment i am trying to think of that word so i forgot what the point i was initially trying to make oh yeah it was lack of a more extensive vocabulary ok thanks peace yeah yeah yeah right

  • @timothyclemson
    @timothyclemson ปีที่แล้ว

    This video is not a meme that is the best way to get a job description for the first time I have a great day and I have a great day ❤️

  • @Roxor128
    @Roxor128 4 ปีที่แล้ว +5

    The "on the phone screen" sentence reminded me of an anecdote I heard about where a program was fed some text containing the word "banana" and got stuck producing "banananananananananananananananana". I think it may have been a variant on Dissociated Press.

  • @EdeYOlorDSZs
    @EdeYOlorDSZs ปีที่แล้ว

    Great explination, I'm defo gonna read that paper

  • @henkjekel4081
    @henkjekel4081 4 หลายเดือนก่อน

    Transformers are also good at Vision tasks, it just wasnt discovered yet how when this video was made. It envolves cutting up images and inputting the patches as tokens.

  • @thatguywhowouldnotsharehis2062
    @thatguywhowouldnotsharehis2062 4 ปีที่แล้ว +48

    This is how my brain works:
    If presenter == "rob miles":
    Views=views+1

    • @oluchukwuokafor7729
      @oluchukwuokafor7729 4 ปีที่แล้ว +18

      error: Views not defined

    • @jbko6161
      @jbko6161 4 ปีที่แล้ว +3

      or Views++
      depends on language

    • @threeMetreJim
      @threeMetreJim 4 ปีที่แล้ว

      Very simplistic, hard coded and no way to learn...

    • @ZenoDovahkiin
      @ZenoDovahkiin 4 ปีที่แล้ว +1

      >Views=views+1
      You seem to be referencing Pythin. Python is case sensitive.

    • @HypnosisLessons
      @HypnosisLessons 4 ปีที่แล้ว

      @@jbko6161 we dont do that here.

  • @juuse94
    @juuse94 4 ปีที่แล้ว +4

    The phone text prediction is based on the user's own inputs. Probably not a neural net still, but they learn to form sentences like you write them.

  • @sevdev9844
    @sevdev9844 3 ปีที่แล้ว

    If I want to have a system on my own computer, not a lot of instances on a server, would the attention based model still be better than a transformer? Difficulty/Ressources/other

  • @anandmoon5701
    @anandmoon5701 4 ปีที่แล้ว +4

    What if we start writing a Novel or Poem or Lyrics dose this transformers influence the way we think.

  • @amir-ali8850
    @amir-ali8850 ปีที่แล้ว +1

    I wish someone would love me as much I love compuerphile videos!

  • @Arnaz87
    @Arnaz87 4 ปีที่แล้ว +1

    Are you teasing us with the data processing video Sean? This is the second video in which it appears at the end but it doesn't exist.

  • @oliverdowning1543
    @oliverdowning1543 3 ปีที่แล้ว

    So does attention remember where to look at rather than what you’re looking at so you just have to remember the words’ “addresses” in the text? Is that right?

  • @cavalrycome
    @cavalrycome 4 ปีที่แล้ว +6

    The "Attention is all you need" paper is available for download here: arxiv.org/pdf/1706.03762.pdf

  • @STEMqueer
    @STEMqueer 4 ปีที่แล้ว +2

    You where right on the initialization with zeros (checked my code). Please read some of the text examples on gpt-2 paper, the unicorns ones it's so credible.
    Good video, congrats!

  • @cartossin
    @cartossin ปีที่แล้ว +3

    It's shocking to see the tone and context of language model conversation change so much in 3 years. Back then it was "well we try to figure out the next word by..." and now it's "the model is simulating a dumb person so it outputs dumb things, but sometimes..."

  • @CSniper
    @CSniper 3 หลายเดือนก่อน

    Would be nice with a deepdive on transformers with Miles!

  • @RichoRosai
    @RichoRosai ปีที่แล้ว +4

    As a translator recently fallen on hard times, I suddenly feel the need to send a robot back in the past to disrupt progress in this field.

  • @_HarshVerma
    @_HarshVerma 2 ปีที่แล้ว +1

    i think it ' s interesting . i ' m not sure i ' ve ever heard of an ai language model - GPT J 6B which i am using currently using

  • @kiffeeify
    @kiffeeify ปีที่แล้ว +2

    13:00 The Irony: Now we have transformers that are stupidly good, but they are not recurrent, which is why they are still limited to a certain max token length. I can't wait to see the combination of a GPT like model with a LSTM like mechanism, where the Transformer keeps a compressed version the history of its contexts, and decides what is relevant for the future, and what he can forget.
    Maybe Attention isn't all you need after all ;D

  • @World_Theory
    @World_Theory 4 ปีที่แล้ว +1

    I suspect that image recognition and language related tasks are more similar problems than people realize, due to the context problem. Context is necessary for both. But what I wonder is: Is context a memory problem for image tasks, like it is for language tasks? Is it hidden because the data described by images isn't in sequence like a sentence?
    I'm thinking of context like the stuff outside the view of a window, which isn't being looked at by a neural network. But now I wonder if context and learned human assumptions are being mixed up with each other some.

    • @TheSam1902
      @TheSam1902 4 ปีที่แล้ว +1

      Image and text processing are alike in the sense that you ought to deduct context BUT a text is already a very densely packed semantic information medium, whilst an image isn't. Just compare the entire work of Shakespeare in text format vs TIFF format. The former can be stored on less than a megabyte whilst the latter takes up more that a hundred gigabytes.
      Therefore the first step is to extract meaningful "features", that's why CNN are so powerful. Though they're not designed to do the job of transformers and attention, they should be traded for transformers in the deepest (topmost) layers IMO.

    • @World_Theory
      @World_Theory 4 ปีที่แล้ว

      Samuel Prevost,
      I now wonder how someone or some thing, decides what is meaningful. Philosophy…

  • @MaxDiscere
    @MaxDiscere 4 ปีที่แล้ว +4

    So and how do I implement it? Some examples would've been great on how to use these models

    • @Guztav1337
      @Guztav1337 4 ปีที่แล้ว

      I would look up the paper. This video has about the same information as the companies would give to the journalists, ie. it is not technical just overview.
      So either search up the model or read the paper.

    • @abcdxx1059
      @abcdxx1059 4 ปีที่แล้ว

      lol they are really hard

    • @abcdxx1059
      @abcdxx1059 4 ปีที่แล้ว

      @@Guztav1337 there are a lot of technical videos on it

    • @Guztav1337
      @Guztav1337 4 ปีที่แล้ว

      Yes, use youtube or other media to find what you are looking for.

  • @shakirsofi877
    @shakirsofi877 2 ปีที่แล้ว

    Amazing !

  • @suleimanmustafa1473
    @suleimanmustafa1473 4 ปีที่แล้ว +6

    I usually pronounce WALK as WORK and THREE as TREE

  • @thelostlinuxlawngnome5437
    @thelostlinuxlawngnome5437 ปีที่แล้ว

    Great video

  • @zekeluther
    @zekeluther 3 ปีที่แล้ว

    Would like to hear about Performer..

  • @letMeSayThatInIrish
    @letMeSayThatInIrish 4 ปีที่แล้ว +1

    Comment from my phone: "The first thing you can say about the app store in a new car in a new location in a car that has a lot to be is the car and a new vehicle that you have a lot to get in a new vehicle to be sure that you can use it to get a better car than a new vehicle and if it doesn't have the right amount you will have a good chance to be a new vehicle".

  • @nickbeaumont2601
    @nickbeaumont2601 ปีที่แล้ว

    lol at all the festival wristbands he's keeping

  • @yassinebarhoumi6973
    @yassinebarhoumi6973 3 ปีที่แล้ว +1

    19:56 now you know with GPT-3!

  • @Micetticat
    @Micetticat 4 ปีที่แล้ว +3

    Are the English subtitles of this video autogenerated? In that case we have an example of a very good language model in action!

  • @Desmaad
    @Desmaad 4 ปีที่แล้ว

    3:00 I think those text prediction systems are based on probability-based n-ary trees.

  • @saurav1096
    @saurav1096 ปีที่แล้ว

    I like to think the AI was drooling over Arnolds arms and makes silly mistakes like , thinking a mugs a dumbbell 😂 17:00

  • @euclideanplane
    @euclideanplane 4 ปีที่แล้ว +1

    2:52 ok but, on top of this, we have our own concepts of what people might say next, in our own minds, based on what the conversation is about, based on what the person typically likes to talk about, based on where you are or what you've been doing that day, or even things you've done with that person in the past like some epic fortnite game you played a week ago he still likes to bring up when you talk about getting a new computer (maybe he thinks he will be even better than he was during that epic game you two played but once he has 244hz monitor and new pc things, haha)

    • @euclideanplane
      @euclideanplane 4 ปีที่แล้ว +1

      3:44 breh, what's the most common word to follow after a combination of TWO specific words? Or Three of Four of Five or Six.
      Does the complexity increase a requirement for more computing power for this to even be useful? I imagine it does, but i'm not sure how many common word stitching we could find, maybe we run it once and we just jot down all the really common ones. Or run it constantly... like our brains do... memory... and constantly remember or generate your own, unique way of stitching words together, one that seems smarter or more detail dense or ... k thinking mode off, you ever just sit and wonder how our mind programs work...

    • @euclideanplane
      @euclideanplane 4 ปีที่แล้ว

      "Doesn't look back"
      Easy way to put it.

    • @euclideanplane
      @euclideanplane 4 ปีที่แล้ว

      9:30 well, seems I understood / understand these things, I'm happy about that

  • @taylorchu
    @taylorchu 4 ปีที่แล้ว +7

    I watched this video on my phone on my phone on my phone on my phone on my phone

  • @rikuown
    @rikuown 4 ปีที่แล้ว +28

    I know transformers, I've seen all the movies.

    • @mj6463
      @mj6463 3 ปีที่แล้ว

      Imagine if all the transformers’ speech was written by the unicorn ai, I’d watch that in a heart beat

  • @nestycb6702
    @nestycb6702 4 ปีที่แล้ว

    I think a great problem about these models is that they rely on morphology. After "say", the two main possible words that may appear are only words with gramatical meaning like "to, with..." or a personal name. This predicts how small groups of words may work together, but is clueless about the overall structure of a sentence. Generative syntax is making a lot of progress in this issue, thus making it possible to program how the structure of language itself works in each language. Although, let's not forget, even past said problem, that would be not enough as it would bring out the importance of function in language.

  • @jmw1500
    @jmw1500 ปีที่แล้ว +1

    Oh I remember this. They were scared of it then too.

  • @tuluwa
    @tuluwa 3 ปีที่แล้ว

    Yea it all makes sense

  • @lukeusername
    @lukeusername 4 ปีที่แล้ว +13

    I am in the morning but I can't find it on the phone screen on the phone screen on the phone screen on the phone screen on the phone screen on the phone screen on the phone screen on the phone screen on the phone screen...

  • @pgoeds7420
    @pgoeds7420 4 ปีที่แล้ว +1

    Today I am going to tell you to respect the Tech Lead. I am the Tech Lead and I am drinking coffee. This is quite nice coffee from the place in the comments but they are not going to give you the discount I get because I am the Tech Lead.

  • @ulissemini5492
    @ulissemini5492 3 ปีที่แล้ว +7

    whos here after GPT-3 becomes self aware?

  • @lakloplak
    @lakloplak 4 ปีที่แล้ว +2

    Im watching now at the 9 min mark and just noticing why I forget stuff and only remember the important bits. If the AI knows what words are important it "knows" what words don;t matter. In a sense: Making AI forget makes it smarter. Hence, I forget a lot so I am smart.

  • @charlieangkor8649
    @charlieangkor8649 4 ปีที่แล้ว

    please provide links to all fulltext PDF papers relevant to your talk. I download them, strip headers and footers, run through pdftotext, clean up and add to my English corpus for teaching my neural network. Thanks.

  • @piad2102
    @piad2102 2 ปีที่แล้ว +1

    I understand a lot you say, but a lot, not. :) Even though that, your enthuisasm is infective. I love people with such commitment, and your humor is great.

  • @Marina-nt6my
    @Marina-nt6my ปีที่แล้ว +3

    👀 2:45 👀

  • @dustinking2965
    @dustinking2965 4 ปีที่แล้ว +1

    I'm curious how attention works.

  • @sdmarlow3926
    @sdmarlow3926 4 ปีที่แล้ว +9

    Transformers: AI in Disguise.

  • @videooblivion
    @videooblivion หลายเดือนก่อน

    13:05 relevance realization, cc: John Vervaeke

  • @dixie_rekd9601
    @dixie_rekd9601 11 หลายเดือนก่อน

    I've been using novelai over the last few days, it's come a long way since this video, it can write incredibly natural text, great for DND worldbuilding.

  • @testeeduandsx5957
    @testeeduandsx5957 4 ปีที่แล้ว +1

    chomskys tree would tell which parts to skip and make the beguining of the phrase affect the end.

  • @starlight_garden
    @starlight_garden 11 หลายเดือนก่อน

    0:50 error in subtitle.

  • @RaphaelAnthony
    @RaphaelAnthony ปีที่แล้ว

    Hi, 2023 here, it’s working. Taking over all industries. Now it’s learning how to give the program specific and detailed instructions for best results.

  • @acwaller1000
    @acwaller1000 4 ปีที่แล้ว +7

    AI play chess
    ------>
    Deep Blue
    AI play Go
    -------->
    Alpha Go
    AI write song lyrics
    --------->
    Darude Sandstorm

  • @rentzepopoulos
    @rentzepopoulos 4 ปีที่แล้ว

    I did that (HMM) in my PhD in 1992... How important it is to have the right context :)

  • @aniketdeshmane6569
    @aniketdeshmane6569 4 ปีที่แล้ว

    Is it possible to get the latest flash player is required for video playback is unavailable

  • @tamasdemjen4242
    @tamasdemjen4242 4 ปีที่แล้ว +3

    I think it might be an advantage to think in terms of phrases, instead of just putting words together. If you've learned a foreign language before, you know that the key to fluency is using native speakers' sentences as templates. I would first choose the subject, verb, object triplet, then try to build the rest of the sentence around those using common idioms. However, it becomes significantly more complicated to author an entire article in a coherent way. Also consider what gives speech that unmistakable human quality, such as the use of filler phrases, like "as a matter of fact", "to be honest", "to put it simply", and so on. Then instead of just asking where the train station is, we would say "I was wondering if you could tell me where the train station is." In fact, this is how I learned English, first studying the sentences of native speakers, then varying some of the words a bit, until I became confident enough to say just about anything I wanted. This may be a bit more difficult in Hungarian, Finnish or Estonian, where the word order isn't fixed, and the words are strongly agglutinative. You miss a single suffix, and you miss the whole meaning. But I'm sure they all boil down to common templates and patterns as well.

    • @tamasdemjen4242
      @tamasdemjen4242 4 ปีที่แล้ว

      As others have pointed it out, ideally the neural network should discover these complex rules automatically, without engineers having to manually input them.

  • @abrahamowos
    @abrahamowos ปีที่แล้ว

    What if " on the phone screen" was the EOS token ? 😄

  • @cherubin7th
    @cherubin7th ปีที่แล้ว +2

    This video kind of skipped the transformer part

  • @mosef312
    @mosef312 4 ปีที่แล้ว

    I am a little thing called love to see if you have any questions or concerns please visit the plug-in settings to determine how attachments are handled the situation I'm not sure if you have any questions or concerns please visit the plug-in settings to determine how attachments are handled the situation I'm not sure if you have any questions or concerns please visit the plug-in settings to determine how attachments are handled the...

  • @hjvisagie
    @hjvisagie 4 ปีที่แล้ว

    I would love a lot more than I would love for them but they are the best legs on my hand. 😂