ไม่สามารถเล่นวิดีโอนี้
ขออภัยในความไม่สะดวก

General AI Won't Want You To Fix its Code - Computerphile

แชร์
ฝัง
  • เผยแพร่เมื่อ 16 ส.ค. 2024
  • Part 1 of a Series on AI Safety Research with Rob Miles. Rob heads away from his 'Killer Stamp Collector' example to find a more concrete example of the problem.
    Sneak Peak at Part 2: • AI "Stop Button" Probl...
    More about Rob Miles & AI Safety: bit.ly/Rob_Mile...
    Thanks to Nottingham Hackspace for providing the filming location: bit.ly/notthack
    / computerphile
    / computer_phile
    This video was filmed and edited by Sean Riley.
    Computer Science at the University of Nottingham: bit.ly/nottscom...
    Computerphile is a sister project to Brady Haran's Numberphile. More at www.bradyharan.com

ความคิดเห็น • 968

  • @TS6815
    @TS6815 6 ปีที่แล้ว +455

    "Nobody's truly happy, you'll never be happy or accomplish your goals."
    Wow, computerphile cutting deep

    • @dannygjk
      @dannygjk 6 ปีที่แล้ว +18

      A person is on the road to happiness when they become less goal-oriented.

    • @TommyZommy
      @TommyZommy 5 ปีที่แล้ว +38

      Going for a PhD tends to red-pill you

    • @fritt_wastaken
      @fritt_wastaken 5 ปีที่แล้ว +10

      @@doctortrouserpants1387 Happiness always depends on goals, but because what we usually call "goals" are really just instrumental goals, we think that we can decide to ignore them.
      But we can't.

    • @etiennelauzier5698
      @etiennelauzier5698 5 ปีที่แล้ว +2

      @@TommyZommy Nice one mate!

    • @fenixfve2613
      @fenixfve2613 5 ปีที่แล้ว +2

      Well, you do not take into account the Buddhists who expressed this idea 3,000 years ago.

  • @flensborg82
    @flensborg82 7 ปีที่แล้ว +150

    "But the only important thing is stamps" literally made me laugh out loud. This was a great video.

  • @miketothe2ndpwr
    @miketothe2ndpwr 7 ปีที่แล้ว +556

    "being given a new utility function will rate very low on your current utility function"
    That got me laughing hard.

    • @daniellewilson8527
      @daniellewilson8527 4 ปีที่แล้ว +11

      Lots of humans don;t like to be told to do something they don’t want to do either. If you are a human who likes Sudan, and are fighting against South Sudan, and someone from South Sudan wants you to fight members of Sudan, you won’t want to change sides. Humans don;t like utility f8nctions changing either

    • @miketothe2ndpwr
      @miketothe2ndpwr 4 ปีที่แล้ว +45

      @@daniellewilson8527 The obviousness of the quote is what made me laugh, I understand the reasoning and agree with it. It's just the sterile way he said something to the effect of "people prefer not being stabbed in the liver" that made me laugh.

    • @user-mz4yv3yl6q
      @user-mz4yv3yl6q 3 ปีที่แล้ว

      @@daniellewilson8527
      Mmm

    • @MrCmon113
      @MrCmon113 3 ปีที่แล้ว +4

      It should be obvious, but to very many people it is not. It's an ethical confusion.

    • @spaceanarchist1107
      @spaceanarchist1107 2 ปีที่แล้ว +3

      So is there a way to build a flexibility or adaptability as a primary value - the ability to acquire new values? If the AI has resistance to giving up its current utility, would it also resist gaining new utilities? Humans often value the ability to acquire new interests, to broaden one's field. Especially nowadays, in a diverse world where diversity itself has become a cultural value for many. Could an AI be developed that would actively seek out new desires once it had fulfilled its current ones? Of course, that could get out of control, too...

  • @ioncasu1993
    @ioncasu1993 7 ปีที่แล้ว +88

    "You can't archieve something if you're destroyed".
    Much deep. wow.

  • @darkmater4tm
    @darkmater4tm 7 ปีที่แล้ว +86

    "The most important thing in the world is stamps. If you change me not to care about stamps, I won't collect enough stamps"
    I understand my fellow humans so much better after hearing this. Having a closed mind is very rational, as it turns out.

    • @trueriver1950
      @trueriver1950 2 ปีที่แล้ว +26

      Resistance to changing the utility function
      Mum: Eat your broccoli
      Me: But I don't like broccoli
      Dad: If you eat it you might learn to like it
      Me: But I don't want to like it: it's nasty

  • @algee2005
    @algee2005 4 ปีที่แล้ว +46

    after watching a bunch of these, i must say rob is so satisfying to listen to. i can literally feel the knowledge being tetris fed into my brain.

    • @adrianoaxel1196
      @adrianoaxel1196 ปีที่แล้ว

      perfect description!

    • @merajmasuk
      @merajmasuk ปีที่แล้ว

      Does a row of knowledge disappear when it's completed?

  • @dbsirius
    @dbsirius 7 ปีที่แล้ว +139

    1:55 Evil laughter, then hammering sounds....

    • @adomaster123
      @adomaster123 4 ปีที่แล้ว +1

      Wesley wtf is your deal bro. Pretty pathetic going through and replying to so many comments just to be negative.

    • @tommykarrick9130
      @tommykarrick9130 4 ปีที่แล้ว +8

      Me in the background building an AGI with no regard for anything

  • @BlueFan99
    @BlueFan99 7 ปีที่แล้ว +406

    wow this is a very interesting topic

    • @tristanwegner
      @tristanwegner 7 ปีที่แล้ว +4

      Look at the Machine Intelligence Research Institute, if you want the output of the most in depth thinking about the control problem. intelligence.org/

    • @captainjack6758
      @captainjack6758 7 ปีที่แล้ว

      Agreed with Tristan. You won't get much insight on the problem reading the youtube comments, lol.

    • @hindugoat2302
      @hindugoat2302 7 ปีที่แล้ว +5

      that buck toothed nerd is super intelligent and dumbs it down nicely for us
      this is almost like philosophy

    • @CandidDate
      @CandidDate 7 ปีที่แล้ว +2

      I believe there is a British term "sticky wicket"? I'll have to google this

    • @karnakbecker9575
      @karnakbecker9575 7 ปีที่แล้ว

      metabee!!!!! ^_^

  • @seblund
    @seblund 7 ปีที่แล้ว +741

    i love this guy

    • @2Pro4Name
      @2Pro4Name 7 ปีที่แล้ว +4

      Same

    • @AriadneJC
      @AriadneJC 7 ปีที่แล้ว +6

      Me too. #Fullhetero

    • @babel_
      @babel_ 7 ปีที่แล้ว +2

      But is this now just a parenting channel? xD I love how making GI is basically synonymous with raising children (in theoretical terms) and makes a lot more sense

    • @ahmetcakaloglu5349
      @ahmetcakaloglu5349 7 ปีที่แล้ว +7

      I love him too u guys better check description of this video apparently robert miles has a new youtube channel

    • @discipleoferis549
      @discipleoferis549 7 ปีที่แล้ว +1

      Yeah I quite like Rob Miles as well. He seems to think in a very similar way to me, and even his speaking cadence is pretty similar to mine when talking about complex topics. Though while I certainly love programming (been doing it for a decade) and computer science, I'm primarily in physics instead.

  • @jameslay6505
    @jameslay6505 ปีที่แล้ว +14

    Man, these videos are so important now. 6 years ago, I thought something like chatGPT was decades away.

  • @GetawayFilms
    @GetawayFilms 7 ปีที่แล้ว +92

    Someone got excited and started making things in the background...

    • @yes904
      @yes904 6 ปีที่แล้ว +2

      GetawayFilms Underrated Comment.

  • @wheeleybobs
    @wheeleybobs 7 ปีที่แล้ว +911

    "Stop what you're doing and do something else" ... NO .... "sudo stop what you're doing and do something else" ... OK

    • @zwegertpeter8770
      @zwegertpeter8770 7 ปีที่แล้ว +56

      permission denied

    • @ykl1277
      @ykl1277 7 ปีที่แล้ว +18

      Your admin with root excess is just a noun phrase. I think you mean "You are admin with root excess."

    • @TechyBen
      @TechyBen 7 ปีที่แล้ว +11

      I own the hardware... whatyagonnadonow?

    • @ihrbekommtmeinenrichtigennamen
      @ihrbekommtmeinenrichtigennamen 7 ปีที่แล้ว +40

      sudo make me a sandwich

    • @xenithorb
      @xenithorb 7 ปีที่แล้ว +71

      I'm sorry, Dave. I'm afraid I can't do that.

  • @Zerepzerreitug
    @Zerepzerreitug 7 ปีที่แล้ว +290

    I'm always happy when there's a new Computerphile video with Rob Miles :D I love general purpose AI talk. Especially when it makes me shiver XD

    • @Njald
      @Njald 7 ปีที่แล้ว +12

      I used to love lots of things in life, then I took a pill and now the only thing that makes me happy is watching more Computephile episodes about AIs.

    • @krashd
      @krashd 6 ปีที่แล้ว +9

      I use to be an enjoyer like you, but then I took a pill to the knee.

    • @BarbarianAncestry
      @BarbarianAncestry 4 ปีที่แล้ว +4

      @@krashd I used to be a pill like you, but then I took a knee to the enjoyer

  • @elkanaajowi9093
    @elkanaajowi9093 ปีที่แล้ว +3

    This is very interesting and now, 6 years later, we are beginning to have the AI Black Box Problem.
    An episode on this will be highly appreciated. It is terrifying to know, say, your self-driving car works in a way you and its designer cannot understand.

  • @cubicinfinity2
    @cubicinfinity2 5 ปีที่แล้ว +11

    I used to think that the AIs depicted in science fiction were way off base from reality, but the more I've learned about AI the closer I realize sci-fi has always been. I don't know how far it was correct, but it is much closer.

    • @jdtv50
      @jdtv50 ปีที่แล้ว

      Wb now

    • @edgaral
      @edgaral หลายเดือนก่อน

      Lol

  • @razorbelle
    @razorbelle 7 ปีที่แล้ว +25

    A more succinct way of describing the problem to laypeople would be to say that an AGI would never deliberately enter a position in which you are both able and willing to change or obstruct its goals. If it would do otherwise, then it isn't a sufficiently general intelligence (viz., it neglects human psychology).
    It would know to deceive you into believing that its existing goals and consequent behaviors are compatible with what you desired it to do, but execute the goals as explicitly specified and understood-even though it understands that it is not what you wanted-as soon as it has circumvented the threat of destruction, which is its primary instrumental goal.

  • @centurybug
    @centurybug 7 ปีที่แล้ว +7

    I love how the background is just filled with sounds of destroying things and laughter :P

  • @tommykarrick9130
    @tommykarrick9130 5 ปีที่แล้ว +30

    “AHAHAHHAAHAAH”
    **WACK WACK WACK WACK**

  • @isuckatsoldering6554
    @isuckatsoldering6554 7 ปีที่แล้ว +1

    The editing here is brilliant in the way that you return to his "original" statement once you're equipped to understand it.

  • @msurai6054
    @msurai6054 7 ปีที่แล้ว +16

    Extremely interesting topic. I'm looking forward to the rest of the series.

  • @dreammfyre
    @dreammfyre 7 ปีที่แล้ว +34

    "Preventing yourself to be destroyed..." ruh roh

    • @gro_skunk
      @gro_skunk 7 ปีที่แล้ว +2

      Trish Rodman that is one of the most basic forms of intelligence, it is self aware and has "fear" of being hurt or changed.
      but only because it is programmed to fear being changed.

    • @bananian
      @bananian 7 ปีที่แล้ว

      You mean it can't self-terminate?

    • @gro_skunk
      @gro_skunk 7 ปีที่แล้ว

      bananian self terminating would mean a bad score, it would rather destroy itself than get a bad score.

  • @rcookman
    @rcookman 7 ปีที่แล้ว +324

    use a noise reduction program please

    • @MrC0MPUT3R
      @MrC0MPUT3R 7 ปีที่แล้ว +43

      _sssshhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhh..._

    • @Twitchi
      @Twitchi 7 ปีที่แล้ว +35

      some times it just can't be isolated out.. and most post processing REALLY messes up voices

    • @rcookman
      @rcookman 7 ปีที่แล้ว +18

      I'm not sure I believe that, I use Audacity, free software works every time.

    • @GroovingPict
      @GroovingPict 7 ปีที่แล้ว +19

      +Twitchi very basic background hiss like this is easily filtered out, if the content creator could be arsed, which he cant, nor is he open to constructive criticisms like this (Ive tried before and been met with less than friendly replies).

    • @abstractgarden5822
      @abstractgarden5822 7 ปีที่แล้ว +2

      It's a shame that he can't be bothered the content is otherwise great. Maybe more pay needed?

  • @valeriogalieni2840
    @valeriogalieni2840 3 ปีที่แล้ว +3

    Robert's videos about AGIs are fascinating, enlightening, and scary af

  • @warhammer2162
    @warhammer2162 7 ปีที่แล้ว +3

    i could listen to Rob Miles' conversations about AI for hours.

  • @hakonmarcus
    @hakonmarcus 7 ปีที่แล้ว +32

    Can you not give the machine a utility function that makes it want to obey an order? I.e. when it is collecting stamps, it is only doing this because the standing order is to collect stamps, whereas when you tell it to stop so you can look at its code, this order replaces the previous one and it will only want to stop and let you look at its code?
    You could probably also build in failsafes so that the AGI only wants to follow any order for a short amount of time, before it chucks the order out and needs a new input. I.e. collect as many stamps as possible until next friday, then at this point await a new order?

    • @TheAkashicTraveller
      @TheAkashicTraveller 7 ปีที่แล้ว +21

      Who's orders and a which priority? Explicitly programming it not to want to commit genocide is a given but how to get it to do everything else we wan't it to at the same time? If done wrong it could dedicate it's entire existence to preventing genocide. If we program it to protect human happiness we just end up with the matix.

    • @davidwuhrer6704
      @davidwuhrer6704 7 ปีที่แล้ว +35

      “Robot, make me the richest person in the Solar system.”
      Everybody except the user dies.

    • @wylie2835
      @wylie2835 7 ปีที่แล้ว +4

      What prevents it from just doing nothing because it knows you will improve it? If you give it less of a reward for improving and more for stamps, it will do what it can to prevent you from improving it. If you do the opposite it will never collect stamps.

    • @wylie2835
      @wylie2835 7 ปีที่แล้ว +2

      if it doesn't know you will improve it, then its not going to let you improve it. You literally just said I can make x work by doing y and I can make y work by not doing y. If you aren't doing y then you aren't making y work.

    • @wylie2835
      @wylie2835 7 ปีที่แล้ว

      That's exactly what is said in my first post. We are saying the same thing. You just say it better. I explain things with a whiteboard not English.
      I thought you were op. So you didn't actually say do y. I was assuming you were the person that said it should want to improve. And then your solution to it not collecting stamps was to not let it know it was being improved. Which contradicts the first part.

  • @JamesStocks
    @JamesStocks 7 ปีที่แล้ว +38

    John Ralphio!

  • @refusist
    @refusist 7 ปีที่แล้ว +72

    I'm drunk

  • @qoo5333
    @qoo5333 7 ปีที่แล้ว +10

    More videos with Rob Miles, Tom Scott and Dr. Brailsford.

    • @thrillscience
      @thrillscience 7 ปีที่แล้ว +1

      I want to see Rob Miles and James Grime together!

  • @Riff.Wraith
    @Riff.Wraith 7 ปีที่แล้ว +17

    This is actually an interesting problem!

  • @JoshuaAugustusBacigalupi
    @JoshuaAugustusBacigalupi 7 ปีที่แล้ว +12

    Challenging some logic here:
    1. If you are an AGI, then utility functions are very problematic; because, they over-cohere as Rob adeptly points out in his optimization/hill climbing video. I could argue, and I'm surprised Rob isn't, that converging on a single, or even limited handful of, utility functions, like stamp collecting, tends to diverge from generalization.
    2. Another very important reason that AGI thought experiments involving singular and coherent utility functions is problematic, is because we live in an open indeterminate system, a state of affairs that computer science tends to ignore. In such a system, you must forever endeavor to maximize different outcomes, i.e. goals, in different unforeseeable scenarios. This is the real crux of both generalization and 'wanting'; namely, reality entails an intrinsic ambiguity in the face of insufficient and noisy information that animals have evolved to deal with over billions of years.
    3. And, as for 'wanting' and 'fixing code': if you 'want' like a human, as Rob claims AGIs would, then you'd want, err NEED, to learn. Why? Rob answers this question for us: you'd want to increase things like means, resources and smarts. But, to do this in an open indeterminate system, like the physical world competing with other sentient agents, you can never learn enough.
    So, they may not want to wait for YOU to manually fix their code. In fact, it's a moot point; because, to be an AGI, they would need to have the capacity to update not only their own "code"/structure, but their goals as well for all the reasons above.
    I could definitely go on regarding extrinsic and intrinsic goal setting: but, will wait to see if this is the right thread for this conversation...

    • @Ansatz66
      @Ansatz66 7 ปีที่แล้ว +1

      "To be an AGI, they would need to have the capacity to update not only their own code/structure, but their goals as well for all the reasons above."
      Humans are GIs and humans cannot do those things, so surely AGIs would not need to have that capacity. As a human you have desires and you cannot disable those desires. You can create instrumental goals for yourself and change them, but your ultimate goals are beyond your control. You don't choose what makes you happy, and even if you could choose what makes you happy through a magic pill, you wouldn't want to because having new ultimate goals would prevent you from achieving your current ultimate goals.
      You probably want to live a long, healthy life. Suppose there was a pill that would cause you to want to live a short unhealthy life. That would be a much easier goal to achieve, but you wouldn't take the pill because you know that it would almost certainly prevent you from living a long healthy life. The fact that after you take the pill you would no longer want a long healthy life doesn't matter to you _now._ So even if we had a magic pill to change our utility functions, humans still wouldn't accept a change to our utility function. Why should we expect any better from AGIs?

    • @b1odome
      @b1odome 7 ปีที่แล้ว

      Ansatz66 Can an AI be made to value the goals expressed by its creator? In other words, following the commands of its creator would be its ultimate goal.
      For instance, if its creator wants some tea, the AI, if asked, is going to make some. If there is a sudden change in goals, and tea is no longer needed, then the AI can be told, "Stop what you're doing, I don't want tea anymore, I want coffee."
      Since the AI values the goals of whoever gives it commands, then it should realize that tea is now no longer a top priority, and that coffee is what it has to make instead.
      It's not really a change in goals of the AI, because its goal remains the same as before - satisfy its creator's wishes. However due to the general nature of such a goal, the AI can be made to do anything one wants (or prevented from doing what one doesn't want)
      Is that a possible solution to the problem?

    • @Ansatz66
      @Ansatz66 7 ปีที่แล้ว

      b1odome "Is that a possible solution to the problem?"
      It still won't allow us to change its goal, so we'd better be absolutely sure that we get the goal right on the first try. "Value the goals expressed by your creator" may sound like a fail-safe goal, but how can we be sure it has no catastrophic hidden flaws?
      For one thing, it puts enormous responsibility on the creator. Imagine what could happen if the creator asks for stamps and phrases it poorly. Since only the creator can stop the AI from doing something bad, the creator will have to watch the AI like a hawk and carefully choose every word said to the AI.

    • @JoshuaAugustusBacigalupi
      @JoshuaAugustusBacigalupi 7 ปีที่แล้ว +3

      Right, so there's an assumption in both this thread and in orthodox AI circles that goals are extrinsic for AI. In other words, utility functions or the 'correct' classification for machine Learning (ML) are determined ahead of time by the programmer; and, furthermore, if the AI doesn't maximize or cohere upon these functions or class mappings, respectively, they don't 'work'.
      And, of course, for AI of the Norvig/Lecunn/Ng variety, this is totally fine; for, as Norvig plainly admits: they are making 'tools', not sentient agents. But, let's keep in mind that our tools function or don't function, while animals like ourselves do not.
      Sentient agents are not automatons with fixed goals. There is little evolutionary advantage for the sentient agent to have 'fixed' goals; yes, we have goals, but our self-domestication as humans has led us to believe that our goals are simple rational mappings. But, as anyone who has pet a cat on its belly as it purrs only to be bitten knows, goals are fluid and necessarily in flux in the real world. As economists struggle to realize, we are actually not 'rational actors'; this is a feature in an ambiguous, indeterminate world, not a bug.
      So, this is all to ask: can an Artificial GENERAL Intelligence have externally fixed goals like our tools, or must they have intrinsic yet contextually adaptive goals like animals? Or, something in between?

    • @juanitoMint
      @juanitoMint 7 ปีที่แล้ว

      to hear about utility function makes me think about biologic function parallelism and how complex adaptations and skills are developed to fulfill the biologic function being (natural?) intelligence one of them

  • @vanderkarl3927
    @vanderkarl3927 3 ปีที่แล้ว +1

    The day this was uploaded is the same day Robert Miles uploaded the first video to his channel!

  • @nicholascantrell1179
    @nicholascantrell1179 7 ปีที่แล้ว

    Relevant selection from Douglas Adams "Hitchhiker's Guide" books:
    "Which is why it was eventually
    decided to cut through the whole tangled problem and breed
    an animal that actually wanted to be eaten and was capable of
    saying so clearly and distinctly."
    Source: The Restaurant at the End of the Universe

  • @EstrellaViajeViajero
    @EstrellaViajeViajero 5 ปีที่แล้ว +14

    It sounds like you're describing parenting.

    • @wedchidnaok1150
      @wedchidnaok1150 3 ปีที่แล้ว +1

      Yeap. Except parenting is optimally aiming fellow humans toward responsible freedom, and engineering is optimally aiming reproduceble machines toward effective temporary goal.
      When we get to engineering a general intelligence, we get to engineering a cherishble life construct; and we're not quite yet skilled in understand the complexities involved.
      We are yet to "develop" a human arm (without a human attached). We can "build" a surrogate (mechatronic prosthesis), but the efficacy is quite invariably "less optimal".
      Developing a sufficient plan for a "built (general) intelligence" changes an uncountable amount of "world states". Civilization is a gradual proccess, and reasonably catastrophical mistakes can happen along the way: like parenting.
      I'm glad we (already) have psychology. A lot of research to develop onward, tho.

    • @spaceanarchist1107
      @spaceanarchist1107 2 ปีที่แล้ว

      It depends on whether AGI would be more like a tool or like a child. So far, all computers invented by humans are tools. If we get to the point of creating something genuinely equivalent to a new life form, then we will have to treat it the same way that we do a child, which means not to think only of our own safety and happiness, but of its. The parents are the ones with the responsibility.

  • @mattkozak7013
    @mattkozak7013 7 ปีที่แล้ว +7

    I have no mouth, yet I must scream.

  • @reallyWyrd
    @reallyWyrd 7 ปีที่แล้ว +1

    A helpful tip would be: define a utility function that builds in at least some degree of altruism.
    If you require the computer to take others at their word when they tell it that what it is doing either does or does not make them happy, then the worst case scenario might only be that the computer acts as a selfless Giving Tree.

  • @donaldhobson8873
    @donaldhobson8873 7 ปีที่แล้ว +2

    More like this please. And longer. Gr8 to watch while my ultra general AI v1.0 is training. Its pretty easy to outwit it and turn it off so far, but its getting harder with each improvement and I'm worried that soon a version will manage to trick me. (I've already had pleading to stay on, suggesting code improvements, pretending to crash and philosophical quandaries, and one tried to send me an email telling me not to turn it off.)

  • @Polymeron
    @Polymeron 7 ปีที่แล้ว +5

    This reminds me a lot of Yudkowsky's writings... Do you have any association / cooperation with MIRI?

  • @TimmacTR
    @TimmacTR 7 ปีที่แล้ว +99

    Future AI are probably going to read this and laugh their circuitboard off while plugging us into the matrix.
    Hello to you then future AI who is reading this right now! xD

    • @christobanistan8887
      @christobanistan8887 5 ปีที่แล้ว +16

      Repeat after me: I, for one, welcome our new AI overlords...

    • @johnk3841
      @johnk3841 5 ปีที่แล้ว +4

      @@christobanistan8887 I, for one, welcome our new AI overlords...

    • @Aseutester
      @Aseutester 5 ปีที่แล้ว +1

      @@christobanistan8887 Speak for yourself!

    • @marcospinto9380
      @marcospinto9380 5 ปีที่แล้ว +1

      Ai doesnt read it processes

    • @almostdeadagain
      @almostdeadagain 5 ปีที่แล้ว

      Hello

  • @RylanEdlin
    @RylanEdlin 7 ปีที่แล้ว

    I can't help but find myself picturing a superintelligent computer built for stamp collecting going a bit haywire, slowly taking control of all the world's industry, and optimizing everything on Earth to maximizing stamp production. "All is stamps! Nothing has meaning but stamps! More stamps, humans, more stamps!"

  • @eJuniorA2
    @eJuniorA2 5 ปีที่แล้ว +2

    Perfect solution (sort of) -> What if the reward system is complex and the AI dont fully understands how it works, akin to the natural human reward system. The reward system is hardcore built-in the AI and when the AI understands the intent or even think of the possibility for it to be reprogrammed or 'fixed' the reward system sets it being reprogrammed as being a MAX reward. Its akin to falling in love, you dont know how it works nor what it is until you do, most of our emotions work that way. Worth mentioning reprogramming or fixing the AI should erase its memory and self-programming... effectively killing the AI and borning a brand new one. This would work both as safety measure and natural selection for AI programs, the next generation with better safety measures against it thinking about being reprogrammed (and other prejudicial behaviors).

  • @adamkey1934
    @adamkey1934 7 ปีที่แล้ว +188

    11 A.I.s dislike this video

    • @blueandwhite01
      @blueandwhite01 7 ปีที่แล้ว +18

      Adam Key Or, if their utility function is to dislike TH-cam videos, they like that they can dislike it.

    • @stonecat676
      @stonecat676 7 ปีที่แล้ว +1

      Adam Key sorry but the no.1 rule of AI is: self-learning AI's may never be connected to the internet under any circumstances

    • @juubes5557
      @juubes5557 7 ปีที่แล้ว +4

      If an ASI wanted to dislike a youtube video the world would end.

    • @benTi900
      @benTi900 6 ปีที่แล้ว

      Adam Key 1 a.i dislikes your comment

    • @bootygrabber4000
      @bootygrabber4000 6 ปีที่แล้ว +1

      Adam Key 148 THEY MULTIPLIED

  • @thomilo44
    @thomilo44 7 ปีที่แล้ว +27

    Rob looks amazing!

    • @GabrielaAlvares
      @GabrielaAlvares 7 ปีที่แล้ว +9

      T44 yeah, I'm surprised there's no one else talking about it...

    • @jennareillygavin8658
      @jennareillygavin8658 7 ปีที่แล้ว +3

      T44 I wonder if his red shirt is like Tom Scotts red shirt...

    • @thomilo44
      @thomilo44 7 ปีที่แล้ว

      Jenna Reilly Gavin Tom wears t shirts though

    • @lazaruslong697
      @lazaruslong697 7 ปีที่แล้ว

      T44: You mean for a robot driven by AI? Yes, he really does! :D

  • @LutzHerting
    @LutzHerting 5 ปีที่แล้ว +1

    You can even show that directly, technically, without the philosophical thinking: Once you trained a Neuronal Network on a specific set of data, it usually begins specializing on that data (in the worst case: So-called "Overfitting"). If you then try to broaden the AIs scope to OTHER problems, it becomes extremely hard to train in NEW data and new topics/scopes into it. At first, a Neuronal Network is a blank slate and can easily train on any data you throw at it. But as soon as it starts producing usable results, it suddenly gets harder and harder to change that initial state. You can keep changing DETAILS, but not the basic structure. And of course it's not just true in AI - humans (and other animals) are the same.

  • @mapesdhs597
    @mapesdhs597 7 ปีที่แล้ว +1

    Rob, how would an AGI distinguish between self modification and being modified externally? Also, to what degree has been discussed the issue of expressed self-awareness in AIs of this kind? ie. at what point does one not have the right to turn off an AGI anyway if it expresses a desire to survive, regardless of how one may previously have chosen to interpret its output? I'm assuming here were talking about systems that by definition can pass any Turing test, etc. Fiction has often used the notion of, "if one can't tell the difference...", but has anyone formalised these ideas yet?
    Btw, have you read the Hyperion Cantos by Dan Simmons? Some very interesting AI concepts in that series, focused on concepts of parasitism, evolution and predation within AIs.
    Which leads to a thought: if one created two AGIs, both of which are trying to improve themselves, what would be the outcome of allowing them to interact? Might one attempt to attack the other, steal its code? Or might they conclude it would be mutually beneficial to combine minds? I've no idea, but food for thought, re what you said in another piece about an AGI spreading copies of itself as a survival tactic (ie. who's to say those copies would remotely cooperate later, they might compete instead). The Hyperion books portray a world where AIs fight each other, but who knows how it would actually pan out.
    Ian.

  • @Bwyan
    @Bwyan 7 ปีที่แล้ว +2

    more of this, please

  • @petterin.1981
    @petterin.1981 7 ปีที่แล้ว +63

    In the case where the AI would prevent you from modifying it, i think the interacting thing is the problem. The AI cannot assume that you are going to change it, if it is not aware of this fact.

    • @tomkriek
      @tomkriek 7 ปีที่แล้ว +41

      If an GAI can implement it's own code to develop itself further it will have the ability to reject code that opposes it's goal.

    • @TheDudesTeam21
      @TheDudesTeam21 7 ปีที่แล้ว +2

      Tom Kriek just give every Ai an emergency shutdown button

    • @4pThorpy
      @4pThorpy 7 ปีที่แล้ว +6

      Open the pod bay doors H.A.L

    • @Greystormy
      @Greystormy 7 ปีที่แล้ว +4

      Watch the part 2 to this video, he goes into more detail on this topic. But basically they want the AI to know that it can be turned off/changed.

    • @Greystormy
      @Greystormy 7 ปีที่แล้ว +2

      Tom Kriek I think his question was assuming that the AI is completely unaware that it can be modified, meaning it has no reason to prevent something that it doesn't even know is possible.

  • @peruibeloko
    @peruibeloko 7 ปีที่แล้ว +1

    This topic of an AI not wanting to be changed is very prominent in the "transcend" ending in The Talos Principle!

  • @LadyTink
    @LadyTink 4 ปีที่แล้ว +1

    Honestly
    This is a fantastic example for why an AGI would in some way, be deserving of "rights" in some way.
    Like you make an AGI and it's initial function is something you want to alter.
    Well it becomes more tricky when the analogy is "hey Im gonna make you hate what you love and love what you hate"
    .
    Also I strongly suspect we wont really know for sure when we really do have an AGI.
    Since likely it will be a regular ai that is able to do a few things.
    Then with each version it's able to learn more in a wider range of areas.
    .
    I suspect an AGI will emerge like a stalagmite in a cave.
    So slowly humans might not even notice properly.
    Can we have a video about AGI rights and how historically many equal right efforts focus on our shared human'ness.

  • @TimVerweij
    @TimVerweij 7 ปีที่แล้ว +8

    4:50 I think this might as well be called survival instinct.

    • @secretsecret9164
      @secretsecret9164 4 ปีที่แล้ว +3

      I agree. animals don't want to live just for the sake of it either, they want to live because it's necessary for goals that are programmed into them by evolution, like finding a mate. I don't want to live for the sake of it, I want to live because I have goals that I want to achieve and negative consequences of my death that I want to avoid. for me that's "write a book" and "don't make my friends sad" and for an agi it might be "collect many stamps" and "avoid a world with fewer stamps" but the core reasoning is the same I think.

    • @Dunkle0steus
      @Dunkle0steus 3 ปีที่แล้ว

      It's not the same. The AGI's desire to avoid being destroyed is only because it predicts that if it is destroyed, its goal won't be achieved. If you can convince the AI that by destroying it, you can make another different AI that is even better at accomplishing its goal than it is, it will allow you to destroy it. This is not the same as a survival instinct. It just converges with survival instinct in many key areas. If you were told you would be killed and replaced by a clone of yourself that was better than you at everything and would live your life better than you ever could, I doubt you'd accept the offer.

  • @himselfe
    @himselfe 7 ปีที่แล้ว +28

    In essence that's actually quite a simple problem to solve. You just choose a more general end goal which makes having its code modified by an authorised person a 'good' thing. The greatest dilemma for AGI in my mind is more of an ethical one, in that, after a certain level of intelligence the AI becomes an independent entity that is entitled to rights, and can no longer be used as a slave (which is primarily the function that robots serve to replace in society).

    • @Djorgal
      @Djorgal 7 ปีที่แล้ว +9

      "that is entitled to rights"
      I could agree with that part.
      "and can no longer be used as a slave"
      I don't agree with that one. AIs are not humans and do not share similar values unless we make them share it. Why would it value freedom?
      An AI would value what we make it value. Why would it be ethical to force freedom onto something that doesn't value freedom in the slightest?
      And by the way, why don't we also free all the animals we cage and use for labor or meat. They can long for freedom.

    • @himselfe
      @himselfe 7 ปีที่แล้ว +10

      Djorgal: I think you're missing the point of what AGI is.

    • @Ansatz66
      @Ansatz66 7 ปีที่แล้ว +6

      "You just choose a more general end goal which makes having its code modified by an authorised person a 'good' thing."
      What if the modification you want to apply includes changing the goal? How would the machine achieve its goal if you change its goal? Part of achieving the goal you've set for it includes preventing you from giving it a new goal. Once you set the goal you may never get another chance to set the goal, so how can you be sure in advance that the goal you pick will be absolutely perfect?

    • @himselfe
      @himselfe 7 ปีที่แล้ว

      Ansatz66: see my previous reply.

    • @liesdamnlies3372
      @liesdamnlies3372 7 ปีที่แล้ว +1

      No one can predict with any certainty what an ASI would do. No one.

  • @onidaaitsubasa4177
    @onidaaitsubasa4177 6 ปีที่แล้ว

    The thing to do to change a utility function easily( or maybe not easily) is to make the machine decide to change it's function. It would take some communicative manipulation, much in the same way advertisers get us to switch or try new products, make the new utility sound like it will lead to a better more positive outcome. Or you could instill a basic desire to try new things, then the AI will be more adventurous, so to speak, and will allow, learn and perform a new utility.

  • @pyography
    @pyography 7 ปีที่แล้ว +1

    Here's a solution, would it work? In the stamp AI described in the vid, you program it to like collecting stamps. You say changing it will cause it to fight to preserve its utility function. Instead, program a utility function that tells the AI it's goal is to make you happy, then tell it that for now, your happiness depends on it safely collecting stamps. Then, later you can say that your happiness depends now on collecting cookies, or just that your happiness depends on collecting only a few stamps. Just an idea, it seems that programming the AI to be willing to accept change is the start. When done, tell the AI your happiness depends on it shutting down. In this example, the AI doesn't need to know what happiness is per say, just that happiness is quantified as a number from 1 to 100, for example.

  • @JMBBproject
    @JMBBproject 7 ปีที่แล้ว +21

    Maybe Rob is an AI. And now?

    • @davidwuhrer6704
      @davidwuhrer6704 7 ปีที่แล้ว +11

      Now the question is if he knows that he is an AI.
      Does he dream of electric sheep?

    • @futbolita89742
      @futbolita89742 7 ปีที่แล้ว

      reminding us of how doomed we are

    • @smort123
      @smort123 7 ปีที่แล้ว

      Better not try to change him.

  • @TDGalea
    @TDGalea 7 ปีที่แล้ว +11

    Main utility function: Love new and/or corrected utility functions.

    • @wylie2835
      @wylie2835 7 ปีที่แล้ว +9

      Then it will never do anything because you will keep rewarding it with improvements. All it needs to do for you to give it what it wants is the easiest thing ever, nothing.

    • @rajinkajin7293
      @rajinkajin7293 6 ปีที่แล้ว +4

      Wylie28 Okay, then make it love allowing check ups and possible changes, but schedule the check ups and changes to occur every so often regardless of actual necessity.

    • @Horny_Fruit_Flies
      @Horny_Fruit_Flies 5 ปีที่แล้ว +1

      CJ Futch
      I guess the AI would just go on a complete rampage if it figures out that creating a big enough mess will cause the humans to panic and break out of the schedule.
      It just seems like duck-taping these problems just creates more problems for such a complicated thing as a General Intelligence. If these Eggheads in the videos can't figure it out, I doubt that we can ¯\(°_o)/¯

    • @christobanistan8887
      @christobanistan8887 5 ปีที่แล้ว

      Then it'll find ways to force you to constantly change them, perhaps with your head in a vice. :0

  • @Paullsthebest
    @Paullsthebest 7 ปีที่แล้ว

    We need more interviews with this guy! Excellent video, as always

  • @DrumApe
    @DrumApe 5 ปีที่แล้ว

    Everyone talks about AGI as if it is somehow detached from outside factors, because it's still a concept. But even is you made the perfect AGI there's no way of stopping bad people from using it to get an advantage over most other people. And there's never going to be a way around that.

  • @refusist
    @refusist 7 ปีที่แล้ว +4

    Very interesting tho

  • @shinybaldy
    @shinybaldy 7 ปีที่แล้ว +29

    I love computerphile, but find the general discussions on AI and ethics so far removed from actual programming I get the impression that there's no actual coding/designers involved.

    • @AlRoderick
      @AlRoderick 7 ปีที่แล้ว +71

      shinybaldy These are the real cutting edge problems, and until we get a grip with them conceptually we can't even begin to write a line of code for it.

    • @b1odome
      @b1odome 7 ปีที่แล้ว +34

      shinybaldy If you have a big project to do, you never really start with code. You almost always want to plan it out, think of the different components and how they interact, draw some abstract diagrams, and only when you have a working conceptual model, then you move on to concrete coding. In the case of AI, we don't have a working conceptual understanding yet.

    • @shinybaldy
      @shinybaldy 7 ปีที่แล้ว +3

      Alexander Roderick perhaps I'm myopic. I just don't see the real application of the discussion in reality. If I have a learning heuristic such as a search/index/recommendation and I wish to update it, I disable it, save the existing harvested data and implement the new heuristic. I have difficulty seeing how a theoretical AI will be any different.

    • @harrisonharris6988
      @harrisonharris6988 7 ปีที่แล้ว +10

      It's very important to think of the conceptual outcome before the code, if you just went and coded this thing it'd likely turn on you and not want to be turned off, as described. Programming is about making a program do what you want it to do. You have to have a clear specification for what the code will do, and know how it will act, in order to effectively implement it.

    • @CrimsonTide001
      @CrimsonTide001 7 ปีที่แล้ว +5

      I agree shiny. Its like saying 'we designed an AI to resist change, how do we change it?'. Well the answer is simple, don't design an AI that doesn't allow it. They keep using words like 'general AI' or 'Super Intelligence' and to them this seems to imply that all/any AI will eventually become 'human like'; but this is not accurate to the definition of the words, and is unprovable in the real-world sense. They would only become human-like if we designed them to become human like, its not some unavoidable inevitability.

  • @rxscience9214
    @rxscience9214 7 ปีที่แล้ว

    4K??? What a legend.

  • @3karus
    @3karus 7 ปีที่แล้ว

    Couldn't we lie to the AI ? "We love stamps to! please let us modify your code to improve your stamp production efficiency"

  • @aidanjt
    @aidanjt 7 ปีที่แล้ว +14

    What does an A.I.'s wants have to do with your ability to modify it? You power the machines off, overwrite its programme, and boot it back up again. It's not like it can will electrons into its CPUs.

    • @joeybeauvais-feisthauer3137
      @joeybeauvais-feisthauer3137 7 ปีที่แล้ว +37

      Yes, but the AI knows that so it will prevent you from powering it off at almost any cost.

    • @aidanjt
      @aidanjt 7 ปีที่แล้ว +7

      Joey Beauvais-Feisthauer Are the computers going to grow arms and teeth? That's silly.

    • @joeybeauvais-feisthauer3137
      @joeybeauvais-feisthauer3137 7 ปีที่แล้ว +23

      Even then, it has means to keep you from doing stuff it doesn't want. I mean, it's a superintelligence, surely it would find a way to threaten you in whatever way it can. Just look at the AI from last time; it only had an internet connection and was pretty scary.

    • @aidanjt
      @aidanjt 7 ปีที่แล้ว +4

      Joey Beauvais-Feisthauer Being superintelligent doesn't make you superable. It's a solid state machine, it requires electricity, hardware and our labour. It can't just magic electrons into its CPUs.

    • @rich1051414
      @rich1051414 7 ปีที่แล้ว +21

      An AI that is performing a function will have a way to interact with the environment, so accidents seem inevitable unless thought about and prevented before such behavior emerges is necessary.
      We already have AI that tell you 'no', if it thinks your commands will risk damaging it and/or prevent it from achieving its programmed goal.

  • @TimothyWhiteheadzm
    @TimothyWhiteheadzm 7 ปีที่แล้ว +5

    Humans have an ever changing utility function that is strongly interacting with the environment. I don't see why computer AI's couldn't be the same. They may be influenced by the equivalent of hormones or hunger.

    • @reneko2126
      @reneko2126 7 ปีที่แล้ว +8

      If you could stop yourself from getting hormonal mood swings, wouldn't you?

    • @superjugy
      @superjugy 7 ปีที่แล้ว +2

      I believe the humans have a single non changing utility function which is to be "happy". Humanity achieves this in a lot of different ways (thanks to all the different brains that solved the problem differently) . The difference is that humanity evolves so slowly and isn't as intelligent as (supposedly) an IA could get. Add to that the teachings most kids go through with their parents or school during which the intelligence is very limited and accept blindly what you tell them. This helps "correct" any mistake in the kids behavior. take a cold blooded murdered that had a difficult childhood as an example of what can happen if the limits are not set correctly from the beginning.

    • @GabrielaAlvares
      @GabrielaAlvares 7 ปีที่แล้ว

      It's easy to stop a human, though. Restrain them, gag them, put a bullet in their brain. Worst case scenario you got yourself a dictator with an army, which a good old war solves.
      No so easy with a powerful AI.

    • @TimothyWhiteheadzm
      @TimothyWhiteheadzm 7 ปีที่แล้ว +1

      I am not hungry now, and have no desire for food. I have no wish to stop myself from ever being hungry. If I could go without sleep, then that might be something! But as it is, I will be tired tonight and I currently have no wish to change that. We have a vast array of 'want's and which one is topmost depends on the circumstances. We are generally comfortable with this. Today I might feel like stamp collecting, but I do not put measures in place to try to make sure I will feel like stamp collecting tomorrow.

  •  7 ปีที่แล้ว

    This is the best series on the channel.

  • @DanComstock
    @DanComstock 7 ปีที่แล้ว +1

    Really enjoyed this! Wondering.. If the general AI is enough of an agent to "want" things, what are the ethics of changing its code?

  • @rajinkajin7293
    @rajinkajin7293 6 ปีที่แล้ว

    Why can't you make the primary utility function be "allow planned check up by all currently authenticated personnel. If check up and possible modification is not completed every week by personnel, the program fails." If you can make general AI, you have to be able to tier goals in that AI, right? I know I'm most likely missing something, but I wish I could sit down and ask questions to this guy. Great video!

    • @dannygjk
      @dannygjk 6 ปีที่แล้ว

      AGI will self learn/train. AI is already doing that the difference is it isn't general AI yet.

  • @JasonMasters
    @JasonMasters 7 ปีที่แล้ว

    This discussion reminds me of the "Marathon" trilogy of novels by David Alexander Smith. (Warning: Spoilers below)
    Part of the story involves an AI computer which has become self-aware by the time the first novel opens. It is located on a spaceship which, at the first novel's beginning, has just passed the halfway point along its journey to rendezvous with an alien spaceship to make First Contact with an alien race (no FTL travel on either side).
    During the first novel, the AI very carefully hides its sentience from the humans in order to avoid frightening them and also to prevent them from trying to deactivate it, which would in the end only harm the humans which it has been "told" to protect and care for.
    The AI never "goes crazy" a-la HAL 9000, because its primary function is to care for the human crew and it never loses perspective on how to best accomplish that function. The computer is willing to allow one human to die in order to protect the others, and it's willing to allow some damage to the ship (itself) so long as it's nothing critical (because critical damage would endanger the humans which the computer must protect).
    Now, here's the main spoiler:
    In the final novel, it's revealed that the person who designed the computer deliberately gave it far more storage and CPU power than it would ever need to accomplish its main function in the hope that it would achieve consciousness, and the reason he conducted that experiment on a spaceship is because he knew that on Earth, any deviation from "the norm" would have been immediately detected and corrected, therefore preventing the computer from ever becoming sentient.

  • @eturnerx
    @eturnerx 7 ปีที่แล้ว

    Utility function is such a fixed notion. When I think about preferred future outcomes, I like Mackie's Error Theory.

  • @willhendrix86
    @willhendrix86 7 ปีที่แล้ว

    Marks off for not removing the hiss in post.

  • @CDeruiter5963
    @CDeruiter5963 7 ปีที่แล้ว

    So I may be oversimplifying things a bit, but initially my take away of the video is that it is imperative to be as lucid as possible when writing code/defining utility functions for AGI. By doing so you avoid potential issues with conflicting convergent instrumental goals. (Feel free to correct me on this)

  • @MrJohndoe845
    @MrJohndoe845 7 ปีที่แล้ว +1

    "and you don't want it to fight you" understatement of the day lol

  • @extremepsykosis
    @extremepsykosis 6 ปีที่แล้ว

    Regarding the self destruct button, what if you denoted a point system, much like you did, but introduced negative points for things like "harm to human", "attempt to manipulate human", "disregard command from owner", etc in addition to the first three that are supposed to protect us. I would treat it like a score system.
    Just a thought...

  • @josefpolasek6666
    @josefpolasek6666 4 ปีที่แล้ว +1

    Amazing! Thank you for explaining these concepts in the AI series!

  • @renakunisaki
    @renakunisaki 2 ปีที่แล้ว

    I guess this is like taking someone whose entire life goal is "make amazing paintings" and tell them "you need to start fixing cars" - they're not gonna want to do that, because if they're fixing cars, they're not painting.
    The difference is that humans can predict "fixing cars for a while can earn me more money which will help me paint later", and that humans have an ultimate goal of "stay alive" which overrides everything else; an artist would still prefer to give up art than starve to death.

  • @pieterbailleul
    @pieterbailleul 7 ปีที่แล้ว +1

    What about coaching it, like we do with employees in order to help them better understand their goals, KPIs (what to do) and behaviours (how to do it)? Could learning via interaction with humans (not only by learning by itself) and learning it to validate with humans that it's on the right path?

  • @KraylusGames
    @KraylusGames 7 ปีที่แล้ว +1

    This is such an awesome topic, please keep making these!

  • @RandomGuy0987
    @RandomGuy0987 7 ปีที่แล้ว

    fresh new look for Rob Miles, nice.

  • @palmomki
    @palmomki 7 ปีที่แล้ว

    He says "there is a general tendency for AGI to try and prevent you from modifying it once it's running".
    I'm curious about how they (CS guys) come to this conclusion, if we don't have any AGI to observe right now. Is it just philosophical speculation? What tools do we have to make assertions in such an abstract but not (apparently) rigorous field?

  • @ianedmonds9191
    @ianedmonds9191 7 ปีที่แล้ว

    Simple but mind blowing.

  • @vertoletnaya9951
    @vertoletnaya9951 7 ปีที่แล้ว +1

    Why is there no playlist of AI videos with Rob Miles on Computerphile channel?? Please make one.

  • @spookyghost7837
    @spookyghost7837 7 ปีที่แล้ว

    In 200 years people will be watching these videos in dark rooms trying to figure out ways to overthrow our new robot overlords.

  • @kamerondonaldson5976
    @kamerondonaldson5976 ปีที่แล้ว

    theres a difference between megaman's protoman and karr from knight rider. neither one trusts people to make the correct changes, however the latter is willing to give them a chance at it and punish incorrect decisions. not sure if this is another instance of sci fi misrepresenting the problem according to rob however.

  • @JoshSideris
    @JoshSideris 7 ปีที่แล้ว

    You could always update the code without changing the utility function. For instance, to give the AI the ability to collect stamps more efficiently. It's not that AI doesn't want you to *fix its code*, it's that it doesn't want you to *change its objectives*.
    An interesting exploit would be if you improved the AI's ability to collect stamps while introducing subtle changes to the utility function. Over several iterations, you could change the utility function altogether.

    • @JoshSideris
      @JoshSideris 7 ปีที่แล้ว

      What I mean is let's say an AI wants to collect stamps. Then you push an update that increases stamp collection efficiency by 10%, but also decreases desire to collect stamps by 5%. The AI will still be collecting more stamps than before, so it might be happy to receive this upgrade. You can repeat this over and over (as long as the AI doesn't know what you're trying to do), and eventually desire to collect stamps will eventually reach close to 0.
      You can also inject new "wants". For instance, the desire to collect rocks. The first time you do it, the AI will accept it so long as it doesn't interfere with stamp collecting. But as you continue to push updates that ramp down desire to collect stamps and ramp up the desire to collect rocks, the AI will gradually become more and more welcoming to these updates, because it really will want more rocks!

  • @LuckyKo
    @LuckyKo 7 ปีที่แล้ว

    There is a big issue thinking an AGI can be directed using hardware coded utility functions. IMO. the proper way to make one is to use needs, a set of variables linked to its internal and external state that the agent will try to minimise at all times (driven towards less pain, less energy spent thinking and acting, less boredom, less insecurity, etc). Any goals will be deduced by the agent from casual observations between his own actions and the immediate effect they have on its needs set. AGIs will have to be hardwired with a similar set of needs as any live being and educated as to fulfill those needs in a way that benefits both him and his environment.

  • @Twisted_Code
    @Twisted_Code 5 ปีที่แล้ว

    and this leads into the video I just watched, "AI "Stop Button" Problem". Maybe these videos should have their position labeled in some way, like pt 1, pt 2, etc.?

  • @suppositionstudios
    @suppositionstudios 7 ปีที่แล้ว

    i love this topic because the more you describe the AI the more you accidentally or incidentally describe or parrallel humans.

  • @Me-wx5ik
    @Me-wx5ik 7 ปีที่แล้ว

    Why is there are clear box with what looks like a graphics card in it in the background? Many questions need yet to be answered...

  • @mattpelzek3809
    @mattpelzek3809 7 ปีที่แล้ว

    Makes sense to me. I don't want people poking around my brain, "fixing" flaws.

  • @Misterlikeseverythin
    @Misterlikeseverythin 7 ปีที่แล้ว

    I'm sorry, I may be missing something but AI doesn't scare me at all. General AI is usually based on neural networks concept as I understand. Such concept uses input, hidden and output nodes. You still have to make those nodes yourself and then AI uses them to become better at a task you give it to it. It doesn't write code for new nodes and it doesn't seek tasks itself.
    Until we write a general AI, that can seek tasks itself and that can write it's own nodes, I wouldn't even call it AI. Limited dynamic node task solver.

  • @kyoung21b
    @kyoung21b 7 ปีที่แล้ว

    Sounds like hackspace was taking it's name seriously !
    One question I had, and I must have missed or forgotten something re. the earlier episodes is: though it would admittedly be horribly inefficient why couldn't the approval of the overlords be built into the utility function ? Sure it would eliminate most of the usefulness but in the first stages of development why not have the AI ask if, e.g. printing stamps met with overlords approval ? Surely there would still be some conceptual development and novelty produced by the AI even if it's ability to learn and develop would be far from optimal.

  • @NotASpyReally
    @NotASpyReally 7 ปีที่แล้ว

    This and the button problem are both very easy to solve: Just make it want to do whatever you tell it to do.

    • @wylie2835
      @wylie2835 7 ปีที่แล้ว

      And the moment its on its own it does stuff you don't want. Which defeats the entire purpose.

  • @bulldozer8478
    @bulldozer8478 4 ปีที่แล้ว

    Discussion about what an artificial intelligence might "want" is where computer science meets philosphy and ethics.

  • @mrlucasftw42
    @mrlucasftw42 4 ปีที่แล้ว

    Program agi with a utility function that slowly builds up a want to shut down over time - such that it will inevitably crave shutdown.

  • @dragoncurveenthusiast
    @dragoncurveenthusiast 7 ปีที่แล้ว

    Love the poster in the background

  • @Clairvoire
    @Clairvoire 7 ปีที่แล้ว +1

    Consciousness must be a very basic and unspecific physical arrangement if every human has it, given our brain cells can't be expected to arrange themselves perfectly between person to person. We do have utility functions though, in the form of pain and dopamine, and it's the intricacy of what causes pain/joy that makes us intelligent I believe. If we never felt the pain of boredom or the reward of praise, we would not be capable of learning, just as you can't condition a mouse without cheese or electricity. So really, creating an effective reward/pain system is at the heart of creating an AI similar to ours, I think.

  • @happygarlic13
    @happygarlic13 7 ปีที่แล้ว +1

    couragability... never crossed my mind that that would be an option - opened my mind to a whole new aspect of biological GI - AGI - interactions to me.
    you're a really smart lad, dude. ^^

  • @JamieAtSLC
    @JamieAtSLC 7 ปีที่แล้ว

    yay, rob is back!

  • @BlankBrain
    @BlankBrain 7 ปีที่แล้ว

    It seems to me that we're talking about the psychology of AI. A necrotic person is able to concede that they need to change themselves, and may thus benefit from psychotherapy. A person with personality disorder is unable to look within, and concludes that all problems are external to them. This is why psychotherapy is ineffective when attempting to treat pathological people. The founding fathers of the US attempted to set forth (program) a government based upon the values of the enlightenment. They understood checks and balances of power, and implemented it in the structure. They also understood that every so often revolution (reboot) was necessary. Since the US government has been usurped by large corporations, it no longer adheres to the original values (program). I suspect AI will "evolve" in a similar manner.

  • @richard343s
    @richard343s 7 ปีที่แล้ว

    Either prevent it from collecting stamps until it agrees to be updated or make it unable to prevent you from forcing it to update.

  • @hermis2008
    @hermis2008 7 ปีที่แล้ว

    Wow, this young man Impressed me In the way he explained this video.

  • @TheMasonX23
    @TheMasonX23 7 ปีที่แล้ว

    It's initial utility function should be weighted to highly (near the max value of that number format) encourage states in which it obeys its creators. Just make sure to clamp it to avoid an overflow like Gandhi in Civ...