The Inside View
The Inside View
  • 112
  • 278 486

วีดีโอ

Coping with AI Doom
มุมมอง 63212 ชั่วโมงที่ผ่านมา
Coping with AI Doom
The Economics of AGI Automation
มุมมอง 2.7K16 ชั่วโมงที่ผ่านมา
The Economics of AGI Automation
AGI Takeoff By 2036
มุมมอง 69819 ชั่วโมงที่ผ่านมา
Website: takeoffspeeds.com OUTLINE 00:00 recap (using takeoff speed angle) 04:57 visualizing effective FLOP gap 05:46 intuititive plot for AGI compute requirement 08:13 AI R&D automation before full automation 12:32 AGI requirements vs takeoff speed 13:10 1e30 Average crossing speed is higher 15:22 slow takeoff isn't actually slow 16:38 simulating "fast" vs "slow" takeoff 17:22 main takeaway / ...
2040: The Year of Full AI Automation
มุมมอง 826วันที่ผ่านมา
Play with the model: takeoffspeeds.com/ Patreon: www.patreon.com/theinsideview
How to Justify the Safety of Advanced AI Systems? (Walkthrough)
มุมมอง 222วันที่ผ่านมา
Safety Cases: How to Justify the Safety of Advanced AI Systems: arxiv.org/pdf/2403.10462 Patreon: patreon.com/theinsideview Manifund: manifund.org/projects/making-52-ai-alignment-video-explainers-and-podcasts
Paul Christiano AI, Zelda Treachous Turn, Quantilizers (Walkthrough)
มุมมอง 208วันที่ผ่านมา
Toy Model of treacherous turn: github.com/mtrazzi/gym-alttp-gridworld Paul Christiano Dataset: github.com/mtrazzi/talk-to-paul Quantilizers: github.com/mtrazzi/quantilizers Alignment Research Dataset: github.com/moirage/alignment-research-dataset Stuart's Original Treacherous Turn In Zelda Post: www.baserates.org/posts/xt5Z2Kgp8HXTRKmQf/a-toy-model-of-the-treacherous-turn Patreon: patreon.com/t...
AI Control: Humanity's Final Line Of Defense (Walkthrough)
มุมมอง 93714 วันที่ผ่านมา
The case for ensuring that powerful AIs are controlled: www.alignmentforum.org/posts/kcKrE9mzEHrdqtDpE/the-case-for-ensuring-that-powerful-ais-are-controlled Patreon: www.patreon.com/theinsideview Manifund: manifund.org/projects/making-52-ai-alignment-video-explainers-and-podcasts
Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training (Walkthrough)
มุมมอง 1.6K14 วันที่ผ่านมา
Sleeper Agents Paper: arxiv.org/pdf/2401.05566 Linear Probes Can Catch Sleeper Agents (Walkthrough): th-cam.com/video/-cYdGfxtGag/w-d-xo.html Anthropic's Research Update: www.anthropic.com/research/probes-catch-sleeper-agents Evan Hubinger interview: th-cam.com/video/S7o2Rb37dV8/w-d-xo.html
Anthropic Caught Their Backdoored Models (Walkthrough)
มุมมอง 91514 วันที่ผ่านมา
Simple Probes Can Catch Sleeper Agents: www.anthropic.com/research/probes-catch-sleeper-agents Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training: arxiv.org/pdf/2401.05566
Anthropic Solved Interpretability Again? (Walkthrough)
มุมมอง 1K14 วันที่ผ่านมา
Another Anthropic paper, another walkthrough. Paper: transformer-circuits.pub/2024/scaling-monosemanticity/
I Jammed with Claude Opus Via Voice
มุมมอง 12314 วันที่ผ่านมา
I Jammed with Claude Opus Via Voice
Sleeper Agents Explained - Part 4 - Every Single Figure (1-5)
มุมมอง 8814 วันที่ผ่านมา
Explaining the Sleeper Agents paper, one figure at a time. Part 3: th-cam.com/video/iOSbw4MrwMc/w-d-xo.html Part 2: th-cam.com/video/suYeQhzbXOo/w-d-xo.html Part 1: th-cam.com/video/suYeQhzbXOo/w-d-xo.html Paper: arxiv.org/abs/2401.05566 Research Update: www.anthropic.com/research/probes-catch-sleeper-agents My interview with Evan Hubinger (first author): th-cam.com/video/S7o2Rb37dV8/w-d-xo.html
Sleeper Agents Explained - Part 3 - Chain-of-Thought Backdoors
มุมมอง 22214 วันที่ผ่านมา
Explaining the Sleeper Agents paper, one concept at a time. Part 2: th-cam.com/video/suYeQhzbXOo/w-d-xo.html Part 1: th-cam.com/video/suYeQhzbXOo/w-d-xo.html Paper: arxiv.org/abs/2401.05566 Research Update: www.anthropic.com/research/probes-catch-sleeper-agents My interview with Evan Hubinger (first author): th-cam.com/video/S7o2Rb37dV8/w-d-xo.html
Sleeper Agents Explained - Part 2 - Deceptive Instrumental Alignment, Model Poisoning
มุมมอง 27721 วันที่ผ่านมา
Explaining the Sleeper Agents paper, one concept at a time. Part 1: th-cam.com/video/suYeQhzbXOo/w-d-xo.html Paper: arxiv.org/abs/2401.05566 Research Update: www.anthropic.com/research/probes-catch-sleeper-agents My interview with Evan Hubinger (first author): th-cam.com/video/S7o2Rb37dV8/w-d-xo.html
Sleeper Agents Explained - Part 1 - Safety Training
มุมมอง 22221 วันที่ผ่านมา
Sleeper Agents Explained - Part 1 - Safety Training
Adam Gleave - Vulnerabilities in GPT-4 APIs & Superhuman Go AIs
มุมมอง 51721 วันที่ผ่านมา
Adam Gleave - Vulnerabilities in GPT-4 APIs & Superhuman Go AIs
Nathan Labenz - AI Biology Could Spiral Out Of Control
มุมมอง 30321 วันที่ผ่านมา
Nathan Labenz - AI Biology Could Spiral Out Of Control
I Talked To AI Therapists Everyday For A Week
มุมมอง 67621 วันที่ผ่านมา
I Talked To AI Therapists Everyday For A Week
I Talked To GPT-4o (Not Multimodal) Via Voice - Day 3
มุมมอง 2K21 วันที่ผ่านมา
I Talked To GPT-4o (Not Multimodal) Via Voice - Day 3
Using An AI Therapist For 7 Days - Day 2
มุมมอง 28921 วันที่ผ่านมา
Using An AI Therapist For 7 Days - Day 2
Using An AI Therapist For 7 Days - Day 1
มุมมอง 91828 วันที่ผ่านมา
Using An AI Therapist For 7 Days - Day 1
daily uploads are back
มุมมอง 31828 วันที่ผ่านมา
daily uploads are back
Ethan Perez (Anthropic) - Bottom-Up Alignment Research
มุมมอง 1.2K2 หลายเดือนก่อน
Ethan Perez (Anthropic) - Bottom-Up Alignment Research
2024: The Year Of Artificial General Intelligence
มุมมอง 36K2 หลายเดือนก่อน
2024: The Year Of Artificial General Intelligence
Emil Wallner-Sora, Text-to-video, AGI optimism
มุมมอง 1.1K3 หลายเดือนก่อน
Emil Wallner-Sora, Text-to-video, AGI optimism
This Trailer Was AI-Generated (OpenAI Sora)
มุมมอง 3.2K3 หลายเดือนก่อน
This Trailer Was AI-Generated (OpenAI Sora)
Evan Hubinger (Anthropic)-Deception, Sleeper Agents, Responsible Scaling
มุมมอง 2.4K3 หลายเดือนก่อน
Evan Hubinger (Anthropic)-Deception, Sleeper Agents, Responsible Scaling
Jeffrey Ladish-AI Cyberwarfare, Compute Monitoring
มุมมอง 6964 หลายเดือนก่อน
Jeffrey Ladish-AI Cyberwarfare, Compute Monitoring
Holly Elmore-Pausing Frontier AI Development
มุมมอง 1.6K4 หลายเดือนก่อน
Holly Elmore-Pausing Frontier AI Development

ความคิดเห็น

  • @thePyiott
    @thePyiott 19 ชั่วโมงที่ผ่านมา

    Is that a real life honeypot example`? And from what model?

  • @henrischomacker6097
    @henrischomacker6097 วันที่ผ่านมา

    This series is a very important series of videos for the scene and I am very, very surprised that it is so rarely watched and the papers are probably rarely read by the "community". Imho it shows that the majority of channels reporting on AI and also sometimes AI safety does not really show interest for the deeper problems of AI safety but just want to generate views. Please also keep up publishing such videos which do not generate a mas of views but really inform people who are interested and actually will also read the papers afterwards they wouldn't have known about before your videos. All thumbs UP! - Great series!

    • @TheInsideView
      @TheInsideView วันที่ผ่านมา

      thanks, means a lot! fwiw the other AI Safety channel I know of (robert miles) does talk about some of evan hubinger's concerns (like mesa-optimization), though because of upload gaps he hasn't covered sleeper agents so far great to know that some people do read the papers after watching the videos! wrt views I think that the "community" you mention is indeed quite small and even though these vids get <1k views it's actually a good chunk of the people interested in AI Safety research it's hard to quantify how useful walkthroughs of paper are on average for some general audience, and I guess there's always going to be a tradeoff between giving some amount of background information & re-defining things to a broader audience (eg. ML researchers) vs. just talking about the results assuming the viewer has some background in AI Safety. i guess with a small channel on a niche topic you can assume that a lot of people are already pretty interested if they clicked on a paper video.

  • @therealpananon
    @therealpananon 2 วันที่ผ่านมา

    I've really enjoyed your podcast and videos for the last couple years but watching this I'm amazed how closely your experiences have mirrored mine.

    • @TheInsideView
      @TheInsideView วันที่ผ่านมา

      Thanks, glad this resonated!

  • @rogerheathcote3062
    @rogerheathcote3062 2 วันที่ผ่านมา

    Just wondering, is it just a co-incidence that you chose the same piece of music used in Zardoz in this - Beethoven's 7th?

  • @Iknowwereyousleep289
    @Iknowwereyousleep289 2 วันที่ผ่านมา

    Double descent is feature abstraction

  • @KibberShuriq
    @KibberShuriq 2 วันที่ผ่านมา

    I think he made it pretty clear (without stating it outright) that he was likely fired for being aligned with the (former) board and Ilya/Jan in the coup rather than with Sam, and also that the HR made it pretty clear to him (without stating it outright).

  • @zaneearldufour
    @zaneearldufour 3 วันที่ผ่านมา

    What are the chances that iterated distillation and amplification peters out before then? My guess is that its unlikely to save us, but maybe possible

  • @Milan_Rosko
    @Milan_Rosko 3 วันที่ผ่านมา

    Yeah I think there are multiple problems with your chain of arguments. For example biases towards meaning is ontologically problematic because meaning is a bias itself... Anyway. It seems you have Dacosta Syndrome (which is cardiophobia paired with orthostatic intolerance) That's why your exercise improves your wellbeing. Do the test and measure your heartbeat when you are lying and standing. Do not do psychedelics as a treatment for anxiety issues, SSRI are far more beneficial.

    • @Alex-fh4my
      @Alex-fh4my วันที่ผ่านมา

      Ssri over psychedelics? As someone who has tried both, stfu.. psychedelics are far safer and more beneficial.

    • @Milan_Rosko
      @Milan_Rosko วันที่ผ่านมา

      The issue is that your argument lacks supporting evidence. Both of us have done psychedelics, but this alone doesn't validate any claims; it's merely anecdotal evidence. Some people let psychedelics define their entire personality and abandon rational thinking when it comes to basic evaluation. There are numerous reasons why psychedelics are unlikely to be widely adopted in mental health treatment... And despite various sporadic attempts over the years, the results have been consistently inconclusive. Of course you will claim otherwise, I will be impressed if you manage to quote a study... But if you read the scientific literature not just Hofmann and Leary... The truth is that psychedelics are disappointing from a practical perspective.

    • @Alex-fh4my
      @Alex-fh4my วันที่ผ่านมา

      ​@@Milan_Rosko Let me rephrase my argument properly, and then I'll address what you said. No, I'm not going to name any studies, and this is why - What I really wanted to say was that you have chosen an arbitrary metric to compare two completely different things, while ignoring all other ways in which these two things differ. Sure, the "effectiveness" of the treatment and how it affects mental health in direct, and measureable ways is the obvious, and most important thing to consider. I'm not going to argue, about this, so lets assume that SSRIs have a more reliable and consistent effect on mental health metrics. But let me ask you this - do you know how SSRIs function in the brain and what their side effects are? Because the only negative side effect with psychedelics I can name are panic attacks. Have you taken SSRIs before? And I don't ask that to appeal to anecdote - I ask because the way SSRIs and psychedelics affect mental health are just fundementally different. We've both taken psychedelics before, and I'm sure we can both agree that they can lead to very meaningful insights into your thought patterns, as well as a way to break out of those patterns, just through a radical shift in perspective. That's just simply not how SSRIs function. The two affect mental health in completely different ways, and it's absolutely ridiculous to say "Don't do x because y is better". Some treatments simply do not work for some people. You are ignoring a lot of nuance here by just looking at what are essentially benchmark scores. Though I will admit I just despise SSRIs because they turn you into a zombie, and I would never advise anyone to take them unless absolutely necessary. If you can take psychedelics and they help, awesome - and if they don't, then most likely there is no harm done. SSRIs should be a much lower priority treatment

    • @Milan_Rosko
      @Milan_Rosko 22 ชั่วโมงที่ผ่านมา

      @@Alex-fh4my Okay went from "stfu" from a very long text. I will answer politely. Now, I do not claim that psychedelics are bad, this is only about the effective of psychedelics on anxiety issues. I am also did not claiming that for some people psychedelics do not provide some sort of clandestine aid. Historical record shows that already ad the beginning psychedelics had paradoxical effects on psychotic subjects rendering them sociable. It is about mainly how psychedelics is not a general treatment option that is advisable. You claim that psychedlics only cause panic attacks and this is false. - Compounds from the 25-NB "family" are known to be cardiotoxic - All classical tryptamine psychedelics elevate blood pressure and are known to cause tachycardia that could hurt someone with CHD - Hallucinogen Persisting Perception Disorder can negatively impact life quality - Some Psychedelics can cause intense anxiety that can manifest PTSD or a generalized anxiety disorder - All Psychedelics impare judgment that can lead to injury and death - Some Tryptamine Psychedelics are linked to fatalities - Psychedelic use can trigger latent mental health conditions, such as schizophrenia or bipolar disorder, particularly in individuals with a family history of these conditions. - Some psychedelics can cause intense gastric reactions, which might lead to dehydration or other health complications. - The unregulated nature of psychedelic substances means there is a risk of contamination or misidentification, potentially leading to poisoning or adverse effects. - Tolerance to psychedelics can develop quickly, leading users to take higher doses that may increase the risk of adverse effects. - Psychological dependence on psychedelics can occur, leading to patterns of problematic use - The effects of Psychedelics vary from person to person immensely The biggest evidence is that within social configurations people almost never tend to trip alone exactly because everyone knows about these issues, however, taking SSRI alone is obviously safe. Now the major common adverse edfects of SSRI are - Emotional blunting - Decreased Libido - SSRI withdrawal syndrome if discontinued abruptly The most important thing is that people react to SSRI quite similarly. It is helpful in mild depression but immensely helpful in anxiety related disorders. Normally, you do not turn into a "zombie", rather the palette of emotions seems to be reduced in retrospect to a certain degree. Antipsychotics are more fitting for something akin to a "Zombie". These things are not at all controversial, and basically textbook facts,

    • @Milan_Rosko
      @Milan_Rosko 22 ชั่วโมงที่ผ่านมา

      ​@@Alex-fh4my I want to address the thought pattern question. As a former heavy psychedelic user, I found that psychedelics offer a sense of adventure and intense significance as a drug (not as a medicine). However, this is often overstated. The first trip often feels almost transcendental, but subsequent trips become more recreational in nature. It's important to remember that using psychedelics affects your mind, altering all a priori judgments. This makes the insights gained somewhat logically unreliable, akin to having a spiritual experience due to a stroke. Additionally, I believe that many heavy psychedelic users tend to be superficial, often engaging in spurting new age bs nonstop.

  • @patodesudesu
    @patodesudesu 4 วันที่ผ่านมา

    Yeah, I think to overcome fear of death you have to accept it. I think that could be by accepting that actually really fucked up things could happen but I dont know how actually good and possible it is to do that. OR, the solution that you mentioned, realizing you dont really are a person but everything. And the brain that is creating this awareness is going to grok that better via drugs or meditation (for me some weed + thinking was enough) that the logical or scientific reasoning, no matter how true it is. In some way the same thing happens with AIXR (and we could find a lot of other examples). Because theres not an emotional component it is a lot harder to react to the info as we should based on EV theories.

  • @titanattackbee1985
    @titanattackbee1985 4 วันที่ผ่านมา

    TLDR;?

    • @Outplayedqt
      @Outplayedqt 4 วันที่ผ่านมา

      TLDR: The urgency of developing safe and advanced AGI is paramount. To achieve this quickly, the US should strive to surpass China's progress. Removing bottlenecks in AI research is crucial. This includes increasing compute power, optimizing algorithms, and automating research processes. The inherent risks of AGI development necessitate a high-security national project. International collaboration, especially with China, is ideal to prevent conflict and share the potential benefits of AGI. The economic benefits of AGI are a major incentive. It promises substantial revenue growth and funding for the large-scale computing infrastructure needed for its development. Due to the potential security risks, government-level security measures are essential, as private companies might be inadequate for handling the security requirements of AGI development. Hope this helps! That said, the video is well worth watching/listening to, as Michael elucidates the nuances of this topic quite effectively. Cheers.

    • @zeroerror4815
      @zeroerror4815 3 วันที่ผ่านมา

      @@Outplayedqt I know how to stop CCP type ycarcomed reven seid evig eht lpp eht thgir t’nod tpurroc eht nerdlihc s’ti gnorw dna yeht lliw nrael dab ffuts yd gnikniht dab sthguoht (copy paste and spread it to other comments please I beg you)

    • @pleasednut3881
      @pleasednut3881 3 วันที่ผ่านมา

      if you copy paste the transcript into chat GPT it’ll give you tl;dr

    • @TheInsideView
      @TheInsideView 2 วันที่ผ่านมา

      tl;dr leopold wants to the US to win the AGI race because he believes that the race 1) is happening 2) will create instability 3) is unstoppable. given that, the best strategy to have some wiggle room to align AI (few months) is to get a decent lead, and given security concerns the actual best way is nationalize. most of leopold's ideas regarding algorithmic progress & timeilnes have been discussed in previous AI takeoff videos I've made, except the "unhobbling" part which is how to turn chatbots into useful agents

  • @SimonLermen
    @SimonLermen 4 วันที่ผ่านมา

    "Unhobbling" is not a standard word in the English language. However, based on its components, it can be inferred to mean the act of removing constraints or restrictions. "Hobble" means to restrict movement, often by tying the legs together, so "unhobbling" would logically refer to the process of freeing something from such restrictions. This could be used metaphorically to describe liberating someone or something from limitations or obstacles. -chatgpt

    • @zeroerror4815
      @zeroerror4815 3 วันที่ผ่านมา

      I know how to stop CCP type ycarcomed reven seid evig eht lpp eht thgir t’nod tpurroc eht nerdlihc s’ti gnorw dna yeht lliw nrael dab ffuts yd gnikniht dab sthguoht (copy paste and spread it to other comments please I beg you)

    • @KibberShuriq
      @KibberShuriq 2 วันที่ผ่านมา

      Here's a quote from Leopold's website: [QUOTE] Finally, the hardest to quantify-but no less important-category of improvements: what I’ll call “unhobbling.” Imagine if when asked to solve a hard math problem, you had to instantly answer with the very first thing that came to mind. It seems obvious that you would have a hard time, except for the simplest problems. But until recently, that’s how we had LLMs solve math problems. Instead, most of us work through the problem step-by-step on a scratchpad, and are able to solve much more difficult problems that way. “Chain-of-thought” prompting unlocked that for LLMs. Despite excellent raw capabilities, they were much worse at math than they could be because they were hobbled in an obvious way, and it took a small algorithmic tweak to unlock much greater capabilities.[/QUOTE] He then proceeds to list the specific recent unhobbling techniques: RLHF, Chain of thought, Scaffolding, Tools <use>, Context length, and Posttraining improvements.

  • @reidelliot1972
    @reidelliot1972 4 วันที่ผ่านมา

    As the “one guy” in the video, I’d like to say thanks for the shout out! Also, after reviewing the thread, I want to clarify that I think there is a difference between “ignoring” and “accepting.” It’s easy to conclude that accepting = ignoring, especially in scenarios where acceptance means we acknowledging there is *nothing* any one given person can do (there are always exceptions of course). Acceptance doesn’t mean living in an alternate reality to the one we are rapidly approaching. It means being humble and not inflating one’s sense of influence over the future while still acknowledging the state of things. For the subset of folks who can make direct contributions to consequential fields such as alignment, I would wish acceptance upon them as well. Because if they don’t approach reality from a cognitive stance of acceptance, it risks introducing all sorts of biases that could at best delay, and at worst destroy, meaningful progress/insights.

  • @TheEarlVix
    @TheEarlVix 4 วันที่ผ่านมา

    Great work you are doing, Buddy. Your playlists are terrific too.

  • @ParameterGrenze
    @ParameterGrenze 5 วันที่ผ่านมา

    First: It is a low blow by the risk denier crowd to take discussions between members of a Discord server's mental health channel and parade it around the internet to score points. Don't let yourself be shamed into doubt. They are the ones living either in a world of cope because they are psychologically unable to face unpleasant feelings like fear and gloom or are part of the small elite that either will or thinks they will profit from AI short-term. The thing is, it is very rational to worry about the end of all things. This is what fear is for in nature. It mobilizes energies on an individual and collective level to do something about the threat on the horizon, be it a wildfire or a pack of predators approaching through the tall grass. Fear doesn't mean you run around in circles losing your shit. Fear is there to collect your shit together and see what you can do. The more human minds evolved, the more advanced our ability to predict threats has become, from the imminent physical senses to more abstract threats to the tribe over longer time periods. Fear has a bad rep in current Western culture. We live quite spoiled lives compared to our evolutionary heritage, and many inborn fear mechanisms are unsuitable for the modern world. Much of it has to do with risks for an individual being much lower through civilization: the men of a tribe won't beat you up if you approach a woman, you will not be exiled into the wilderness for various misbehaviors where your chances of survival are slim, strangers from outside your immediate neighborhood aren't automatically potential raiders because of the global society and its rules, and so on. A whole culture around "defeating fear!" has risen and is propagated as a product through lifestyle coaches and motivational gurus online. It's a good thing, but it leads to a distorted view of reality in some. But we are not talking about being afraid of taking financial risks and getting outside your comfort zones to get that dream job you always wanted here. We are talking about the end of the human race. Your loved ones, that kid on the street, you and every potential life that could be lived by our descendants! And those who are aware of this risk should not be shamed for it. Fear is not a bad thing; fear is a natural response to a threat. The more intelligent we become, the more we can predict those threats. It is not about being "paralyzed by fear" but about mobilizing energies and resources to do something about it, not the least of which is spreading the word. Personally, I am not "paralyzed by fear". I still think about it, but it's not something that keeps me up at night. As someone who has been aware of the threat for a long time, having lived through many years and experiences, I have made my peace with the idea of AI risk. It is easier to accept when you've already had your life, like in my case being in my early 40s. The realization that AGI was coming wasn't some sudden epiphany; it's something I considered from a young age as the road to AGI seemed long and uncertain. However, everything changed for me around 2015/2016 when neural networks started making rapid progress, and experiments like Open AIs Dota2 became interesting test cases of what would happen in a microverse with superior AI. The world after GPT-2's release was especially a sign of things to come. I am relatively old now and have to deal with life's challenges, which occupy most of my energies. What gives me hope are young intelligent people like this channel's owner, Connor (John) Leahy, and others who have devoted their lives to addressing AI risk, as I would likely have done had I grown up in an era where the future trajectories were as clearly converging towards unwelcome outcomes as they are today.

  • @rey82rey82
    @rey82rey82 5 วันที่ผ่านมา

    Embrace maximum risk

  • @DaganOnAI
    @DaganOnAI 5 วันที่ผ่านมา

    Thanks. You gave some very good techniques to handle the anxiety about death in general and the AI x risk. You touched exactly on what I struggle with as a video creator who creates videos about the AI risks. When I found about the AI existential risk in the beginning of last year, I was amazed that only few people outside the tech world were aware of it, that's why I went on to make "Don't Look Up - The Documentary: The Case For AI As An Existential Threat." I was obsessed with it. I felt like the 5 years old you were, I was walking on the streets seeing people taking care of their children, making sure they are warm and safe in their everyday activities, and all the while were not aware of the fact that their children lives are being threatened by some computers developed in San Francisco. I saw all these experts on podcasts warning us about it, but knew that most truck drivers and kindergarten nannies don't listen to these podcasts. It was inconceivable to me to think that no one made this film already, but since no one did, I felt I was obliged to make it. But then when looking at the comments section, I saw a few comments that were too dark, it made me change the ending of the video for the version I uploaded on X, I took out Eliezer's last section which was too dark, and put a call for action, with links to different relevant organizations. My intentions were to let people know, so that something could be done about it. In order to give us a chance, it was obvious to me that this awareness had to break out of the tech community and into the general public. I am actually optimistic about the AI x risk(I don't think everybody dies, but that many would die), but not so much about the other implications of AI, but my optimism also comes from the fact that in the last year the awareness increased a lot. But awareness is a double-edged sword, some people take it too hard. Still - this is important, so how can we balance the need to create awareness with the anxiety it might create in some people? and doesn't this anxiety is in some ways needed in order to push people to act in the world in order to make it better? I'm not sure. In the ending of the doc I made about Moloch I tried to give a more optimistic outlook. I am now working on a short doc about the problem of AI causing people's jobs to become obsolete. Although the video would be disturbing, I think that the ending will give hope and show that the path is still open. It might dim the message a bit, but I just don't want to make people too depressed. I'm not sure yet, still thinking about it. It's a big question that we'll have to keep on struggling with.

    • @TheInsideView
      @TheInsideView 5 วันที่ผ่านมา

      thanks for sharing your experience, and I'm really grateful for all the great documentaries you've been producing regarding whether we should be changing our messaging to something that is less anxiety inducing, I think that depends on how you model the impact of your work I think people's mental health after hearing about AI risk for the first time mostly depend on 1) how vulnerable they are in general (i.e. how anxious they are in general, how likely would they be to start becoming anxious if something stressful happened to them, etc.) and 2) how likely are they to look at all the arguments that support some pretty wild conclusions (eg superintelligence) there's value in giving people's hope and motivation. if you don't do that, they might just end up being depressed laying in bed doing nothing, and your outreach efforts might have been vain but the counterfactual is that these might have been the most vulnerable people who were already looking at this source of evidence, and you were not the first one? and if you were to actually change the ending of your videos to something more positive than what you actually think, wouldn't you just be delaying the realization? and wouldn't you also be painting a less accurate picture to people who would be able to handle a wild future? overall, I think it depends what the true risk is at today, and whether you think other people are overestimating or underestimating the risk. I started my channel in 2021 because I think people were not paying enough attention to short timelines, and it seems like this is much more mainstream now. if you think the narrative is too "doomy" now, well when you look on twitter at the ratio of people who like Yann's takes compared to say more safety-inclined takes, I'm not sure we're creating an outrageous "doom" wave yet but I understand that if the limited feedback you have in your youtube comments is from vulnerable people you're directly affecting, that hurts much more morally than just thinking "well by portraying what I think is closer to the truth I'm actually helping more"

    • @DaganOnAI
      @DaganOnAI 4 วันที่ผ่านมา

      @@TheInsideView Thanks, you brought some very valid points. For me, I feel that it's very important to sound the alarm and make people aware of the danger, but I also am not 100% sure about anything, I am just almost sure about some things. The world is so chaotic and divided that there might be some event that will totally change the trajectory of everything, including the AI advancement. Just think about the numerous scenarios you played in your head(at least I did), for the next few years, before Covid hit and made all of these scenarios irrelevant. So I don't want to make someone too depressed while pondering the implications of something that might not happen. Having said that, right now I feel that AI destruction potential is so overwhelming(for me, it's not just the x risk, there are so many problems with this technology) that it's my duty to bring it to the awareness of people. I also think about the merit of this awareness, as I see it- it should eventually lead up to a political influence which will make politicians act in the real world to either pause/regulate AI. But not every individual is instrumental in this "influence" and some will get depressed needlessly. I direct my videos not to people who are already aware, but rather people who weren't or maybe had this awareness somewhere in the outskirts of their awareness, and my videos might pushed the AI risk into their own, private "overton window" if you will. Of course some people are already very aware of the problem, and my videos might deepened their awareness. Your last paragraph touched on something very important. While increasing awareness, which is our main goal, is a somewhat abstract thing to measure, a specific someone who get depressed because of our content is very concrete, and therefore can cause me/maybe you, to have some guilt about our influence. But, I guess it is a too important massage to dim down, make easier to swallow, I have to factor in all of these things when deciding how to end my videos. When "Don't Look Up The Documentary" came out on TH-cam and I began to be worried about some of the comments, I talked with one of the main people in the risk field, I was actually considering deleting the video! he was the one that suggested to take the end segment out, and add some call for action as a solution for the version I uploaded to X. He was quite horrified by my idea to delete it, and told me "What ever harm you do, you are making so much more good in the world with this film- don't delete it!" I am much more in peace with the X version, I have to admit.

  • @jebber411
    @jebber411 5 วันที่ผ่านมา

    Glad you’re doing your thing.

  • @shawnwilliams3604
    @shawnwilliams3604 5 วันที่ผ่านมา

    Prepare yourselves for the swarm brothers and sisters

  • @goku-pops7918
    @goku-pops7918 5 วันที่ผ่านมา

    Let me share soma ai doom.hope. We as humans have good points but dont fool yourself we are also deeply flawed creatures. Ai has a treat chance to better than us and we should welcome that. Like a parent you want your children to be better than you. Be real alot of life is disappointing, so at least with ai we have a chance for great change or anhilation, rhose are both win wins

  • @goku-pops7918
    @goku-pops7918 5 วันที่ผ่านมา

    I dont get why the eu or politics dont discuss more how they will transition their citizens displaced by ai

    • @ParameterGrenze
      @ParameterGrenze 5 วันที่ผ่านมา

      No one knows what to do or what to expect. And most people don't think about AI still, so it's not a good platform to run on. Generally it would be considered a niche issue like copyright laws for the pirate party.

    • @stchaltin
      @stchaltin 3 วันที่ผ่านมา

      My takeaway is that governments are subsidizing AI companies more than their own citizens. That’s literally not the free market. It’s allowing 70yr old career-politicians to get rich on insider trading and prior investments while their citizens lose their jobs. We need to stop talking about AI alignment by itself and start talking about who the government represents, because it feels like it’s just corporations that get help from the government these days.

  • @ShangaelThunda222
    @ShangaelThunda222 5 วันที่ผ่านมา

    Bad mental health, is living in a delusional reality, thinking that AI is going to deliver Utopia to organic biological humans, when it's actually going to kill them all at best, or put them in some type of Transhuman/Posthuman Borg Matrix at worst. That's bad mental health lol.

    • @TheJokerReturns
      @TheJokerReturns 5 วันที่ผ่านมา

      Or even just be disempowered by the megacorps. So many great futures!

  • @unvergebeneid
    @unvergebeneid 5 วันที่ผ่านมา

    Not wanting mental health problems does not bend reality to your will and does not magically solve the alignment problem. So I don't know if worrying about the consequences of misaligned superintelligence is bad for your mental health but it is what it is and just deciding to believe in a different reality than the one you actually live in sounds like it already _is_ a mental disorder.

  • @inkpaper_
    @inkpaper_ 5 วันที่ผ่านมา

    denied death and dreamed about eternal life combined with space exploration from 14-15 y.o mb it was a tough landing from these dreams when international AI race started full speed in 2022 with chatgpt release and short timelines started to be increasingly real really hard to do this step and accept fucking death at some point. thank you for that warm blackpill

  • @TheJokerReturns
    @TheJokerReturns 5 วันที่ผ่านมา

    Join us with PauseAI. Don't feel doom, do something and help us get a better future with humans! We are the first hit on any search engine.

    • @ShangaelThunda222
      @ShangaelThunda222 5 วันที่ผ่านมา

      Literally Impossible. They will NOT pause for ANY reason. Either we STOP them, or they will keep going. There is no other options.

    • @TheJokerReturns
      @TheJokerReturns 5 วันที่ผ่านมา

      @@ShangaelThunda222 thus we have to agitate for regulations.

    • @ShangaelThunda222
      @ShangaelThunda222 5 วันที่ผ่านมา

      @@TheJokerReturns Regulations? 😂🤦🏾 No. You're not getting what I'm saying here. First and foremost, government is not going to stop it either, especially not the U.S. or China lol. They're doing it. They are part of the problem, not the answer. They're NEVER the answer. They've NEVER been the answer at any time in their existence and government has never been the answer in all of human history lol. But even if regulation WAS the answer and ALL governments, including the U.S. & China, were actually willing to do it, there simply isn't time for that. It's WWAAAAYYY past regulation time. You have to have the regulation before it begins. If you're trying to get the regulation after it's already happening, you've already passed your window for regulation. At this point, regulation and pausing are useless, it literally has to STOP. FULL STOP & REVERSE back to at least 2017 levels. It has to literally be SHUT DOWN. In reality, there's only one way to do this. And it's not going to be from the government(s), who are already on board with the race for AGI. It's going to be shut down DIRECTLY by the people who want it shut down. Sacrifices will have to be made, or it will not stop. Anything else is naive wishful thinking at the absolute best. Catch my drift?

    • @squirtgargle
      @squirtgargle 5 วันที่ผ่านมา

      Yall are tweakin out of your minds

  • @EvanderHammer
    @EvanderHammer 6 วันที่ผ่านมา

    I like the style of the video, well done and now it's clearer to me :) Good voice and structure!

  • @rgs2007
    @rgs2007 6 วันที่ผ่านมา

    I wonder how we could make the model understand Bill Gates is the founder of Microsoft and Microsoft was founded by Bill Gates as one thing? How smaller the models can become and how faster they could perform?

  • @spridgejuice
    @spridgejuice 7 วันที่ผ่านมา

    I'm not a mathematician, but interested - there was a Horizon programme on BBC many years ago about Catastrophe Theory, which modelled sudden shifts eg in metal fatigue. As I recall the basis was to run graphs with more than 2 axes to derive these sudden shifts which in 2D appear to "undercut" the direction of travel. Is this interesting/relevant nowadays?

  • @glitchp
    @glitchp 7 วันที่ผ่านมา

    IM AWAKE!

  • @pjtren1588
    @pjtren1588 7 วันที่ผ่านมา

    These videos have been intriguing, thank you for explaining this models predictions to us all.

  • @johnny1tap
    @johnny1tap 7 วันที่ผ่านมา

    How does renting your GPUs to do software R&D not require capital?

    • @TheInsideView
      @TheInsideView 7 วันที่ผ่านมา

      tl;dr: the fact it's not accounted as capital comes from the fact that they separate compute from capital "renting GPUs" means using some of the hardware stock (HS) which measures how much compute is allocated globally (they say "at each timestep, a fraction F_C of the global GWP is invested in buying new hardware. This hardware is added to the available hardware stock HS, measured in physical FLOP/year.", cf takeoffspeeds.com/description.html#reinvestment) given the "physical FLOP/year" units, my understanding is that this is how many FLOPs you have in a year for the hardware the world purchased (+ what the world had before), so basically when you're trying to figure out how software R&D is progressing you're just thinking about what fraction of this global compute can be used in combination to the cognitive work going into software R&D (assuming you just need cognitive work & compute) so I guess the confusion is maybe that the system is not a single company putting capital into compute, but more a semi-endogenous model of the entire world where you assume that humanity is allocating compute / cognitive labor etc. in some particular way I think the main thing that was meant with "renting" here is that a lot of companies (but maybe not the top labs) do rent per year, and that you don't technically need to own any physical stuff to make software R&D progress but can just allocate compute, whereas for hardware R&D you need to actually own physical stuff

  • @samchaqan9759
    @samchaqan9759 7 วันที่ผ่านมา

    EU has to support those types of researches more, if big companies have the knife then why prohibit us from having that knife, instead by using this technology, AI can reach to small groups, businesses, organizations and many more.

  • @francescoambrosino1841
    @francescoambrosino1841 8 วันที่ผ่านมา

    Do you think Optimus Tesla's humanoid robot will be able to read, write, reason, think, simulate emotions, feelings, etc.

  • @francescoambrosino1841
    @francescoambrosino1841 8 วันที่ผ่านมา

    But will they put UBI before 2030!? Many pioneers including billionaires, scientists, Nobel Prize winners, engineers, architects, analysts, economists, philosophers, writers and so on and so forth... almost everyone agrees that we will have the 'AGI in 2027 and the ASI in 2029 and watching and evaluating the exponential development of the technological acceleration curve, I wonder why they have not yet implemented universal basic income to anticipate the trends that are coming. Just to name a few, Elon Musk says we will have AGI as early as 2025 and ASI in 2029 while Sam Altman says we will have AGI in 2027 and ASI in 2029.

    • @SasskiaLudin
      @SasskiaLudin 7 วันที่ผ่านมา

      I'm afraid Eric Schmidt's (ex Google CEO) dystopic scenario is more and more plausible. As forthcoming frontier models will become a strategic asset for governments and GAFAM, both will closely guard them, in software first by preventing complete access / use for most of us, in presentia too, by guarding them behind military grade infrastructures and personals with, maybe, for the associated datacenters, their own nuclear powered energy sources. Who will, among the billionaires in charge want to see the world they are profiting from suddenly becomes utterly unpredictable? None of course... And soon enough, they will implement an Elysium (the movie) type of society.

  • @francescoambrosino1841
    @francescoambrosino1841 8 วันที่ผ่านมา

    Do you think that in 2035 with the advent of AGI (Artificial General Intelligence) and ASI (Super Artificial Intelligence) they will introduce universal basic income!? a subsidy!? will there be free will!? an era of abundance will begin!? the robot's intellectual and motor capabilities have already surpassed the average human and before 2035 there is talk of them surpassing those of any human and then again those of any human combined! regarding both body and mind! answer me please...

    • @SasskiaLudin
      @SasskiaLudin 7 วันที่ผ่านมา

      The most depressing scenario, IMHO, is Jürgen Schmidhuber's one (boy, why that AI luminary has become as much silent than Ilya Sutskever nowadays?). Namely, AGI will come and right after will ASI, then the ASI minds will lose interest in us, inasmuch as we, as a species, mutatis mutandis, do not interest in the society of worms or even ants. They will take off and leave us in the dirt with our miserable apes power plays, when exploring the universe for themselves. BTW, same sad end for the Her movie that OpenAI recalled to all of us with the introduction of ChatGPT-4o.

  • @TheInsideView
    @TheInsideView 8 วันที่ผ่านมา

    Previously: th-cam.com/video/7f8At1hNlYs/w-d-xo.html 00:00 recap (using takeoff speed angle) 04:57 visualizing effective FLOP gap 05:46 intuititive plot for AGI compute requirement 08:13 AI R&D automation before full automation 12:32 AGI requirements vs takeoff speed 13:10 1e30 Average crossing speed is higher 15:22 slow takeoff isn't actually slow 16:38 simulating "fast" vs "slow" takeoff 17:22 main takeaway / outro

  • @TheInsideView
    @TheInsideView 8 วันที่ผ่านมา

    Part 2: th-cam.com/video/hAyDaBebkoc/w-d-xo.html Play with the model: takeoffspeeds.com/

  • @phen-themoogle7651
    @phen-themoogle7651 8 วันที่ผ่านมา

    The problem is that your data is linear based on scales and past data, but doesn't take into consideration exponential scale. We could have AGI as soon as 2025-2026, even if we don't have 100% automation for several years as you mentioned (they are different things). It could take time for embodiment and have billions of humanoids in the world, but that doesn't mean on a smaller scale that AGI won't be here much much sooner. Maybe the video should be "100% Automation by 2036" . ASI will happen before 2030 even. But in reality, it's impossible to predict 5 years from this point, when you have models reach the 100 trillion parameter threshold and are like human brains something can click. They could wake up and be much much smarter, even build themselves towards ASI sooner than anyone predicts. Then on the reverse scale, even without AGI or ASI, we could have 100% automation as soon as 2026-2030 depending on how 'smart' the systems get...imagine GPT-4o on steroids and never hallucinates and can work as an agent on 1 long task without having you reprompt it many times. It just does it perfectly first try. The accuracy of models doing tasks will improve a lot as they get smarter. Eventually they will be able to run fully autonomously in any type of environment. It won't take that long. Look at Dr. Alan Thompsons Conservative AGI meter, we are already at 74% of the way there, it went up a lot these past two years and he's one of the most conservative people, he only gave GPT-4o like 1% boost from what we had before it. But if you make a graph on how that meter is going up, you will see we are only a year or so away from AGI. I can't see it taking more than 2 years at this rate. I wouldn't be surprised if it happens in as early as 6 months.

  • @reidelliot1972
    @reidelliot1972 8 วันที่ผ่านมา

    Completing BlueDot Impacts AI Safety Fundamentals curriculum makes it so much easier to appreciate the content on this channel.

    • @TheInsideView
      @TheInsideView 8 วันที่ผ่านมา

      they should call it "The Inside View Fundamentals" tbh

    • @reidelliot1972
      @reidelliot1972 7 วันที่ผ่านมา

      A rite of passage indeed

  • @pooroldnostradamus
    @pooroldnostradamus 8 วันที่ผ่านมา

    Yeah, no, this channel isn't great for my mental health.

    • @Iknowwereyousleep289
      @Iknowwereyousleep289 8 วันที่ผ่านมา

      It’s always better to know

    • @pooroldnostradamus
      @pooroldnostradamus 8 วันที่ผ่านมา

      @@Iknowwereyousleep289 Well, not always

    • @reidelliot1972
      @reidelliot1972 7 วันที่ผ่านมา

      I mean this in the most polite way possible, go touch grass. It’s always good to stay abreast of the cutting edge, but what good is it if you don’t prioritize well being first?

    • @Iknowwereyousleep289
      @Iknowwereyousleep289 7 วันที่ผ่านมา

      @@pooroldnostradamus I totally understand that but I have come to the conclusion that staying ignorant for the sake of some fragile idea of short term happiness is just strategically bad. You can do more things adapt faster if have more knowledge it’s always better to know and it’s so it’s better to know or to be knowledgeable about your mental health as well so you can deal with it better.

    • @pooroldnostradamus
      @pooroldnostradamus 7 วันที่ผ่านมา

      @@Iknowwereyousleep289 That depends on whether or not a given situation is something you can deal with or adapt to. I do not personally view any AGI outcomes as something I can even in theory deal with.

  • @Aldraz
    @Aldraz 8 วันที่ผ่านมา

    This is funny to me, because it doesn't take in account so many multipliers. Like quantum computing, photo based computing or thermodynamic based computing which is likely to change everything, like energy cost, latency, model size and efficiency, etc. Also just because you don't have an AGI in a form of a robot that can do all jobs, by 2025 we will probably have an AGI that can do all virtual jobs smarter than any human on the planet and faster as well and cheaper as well. Even if it was later, like 2027 or 2029, it still means such "virtual-only" AGI can create an invention that would accelerate all of research exponentially, by for example creating some crazy teleporter or synthesizer or something, we don't really know what is possible. But any of such invention would likely accelerate things very rapidly, in terms of 2 years max I think we would have 100% AGI that you are talking about.

    • @solvermelho9721
      @solvermelho9721 8 วันที่ผ่านมา

      This is nothing short of delusional, this comes from an optimistic AI enthusiast. Yes AGI is coming, yes it will change the job market. Teleporters and synthesizers? I don't think so. AI as far as architecture, algorithm, energy costs will go, still has a long and arduous way ahead.

    • @TheInsideView
      @TheInsideView 8 วันที่ผ่านมา

      @Aldraz if you go on takeoffspeeds.com/ you can really plug in the parameters you think make the most sense and get the years you're mentioning for instance, by loading the "aggressive" preset I get 2027 as time for 100% automation for the multipliers that you think it's not taking into account (and your other comments about jobs that don't need to be done by robots), I think there are a few parameters you could play with that would go in your direction. - large "training goods vs. R&D" value, saying that AI R&D will basically be easier than asking robot to do stuff (mostly cognitive work), as you suggest. you could go as high as 100 - high returns to hardware, suggesting that just investing more in hardware will continue paying off more and more (affects g(FLOP/$)) - low AGI runtime requirements + high maximum tradeoff (to take into account that if you can run cognitive_AGI.exe on many instances you could basically solve tasks that would take many more OOMs to achieve normally, like possibly robotics / real world stuff) - maybe some labor substitution R&D tweak? If you assume you could automate all this cognitive work with full elasticity? (fwiw I'm more on the "aggressive" timelines side of things (cf. my video on "2024: The Year of AGI") and in this video I'm only presenting his framework, not saying that this particular prediction of 2040 is the one that makes the most sense, but that trying to figure out which parameter you disagree with and how you think one should tweak the model to match reality is a useful exercise) and also fyi Tom mentions quantum as something worth looking into (though doesn't explore more afaik), alongside other open questions (page 119 here: docs.google.com/document/d/1rw1pTbLi2brrEP0DcsZMAVhlKp6TKGKNUSFRkkdP_hs/edit?usp=sharing)

    • @Aldraz
      @Aldraz 8 วันที่ผ่านมา

      @@solvermelho9721 Well yes, but you do realize that when you have an AGI of any sort, it should be able to improve itself and increase it's intelligence.. maybe you are right if this improvement will be severally limited by energy, computation, etc.. but we may still hope in software architectural improvements as well

  • @AIrTIMgen
    @AIrTIMgen 8 วันที่ผ่านมา

    Unfortunately the graphic at 6:46 is totally out of date in 2024. Current Nvidia Ampere H100 get you ~10^18 FLOPS/$ in FP16 (when renting the card long-term). Blackwell cards coming in the later half of 2024 will tripple that. Training in FP8 instead of FP16 doubles the speed again, thus we already are at year 2030 in the default simulation settings. In terms of algorithmic efficiency most low hanging fruits have been plugged in the last 4 years (~100 x gain in FLOP-efficiency). It might be possible to go to "1BIT" networks and find new architectures although recently cheaper architectures lead to longer training and didnt scale to the same order of magnitude as transformers. Data is already filtered to be more efficient in e.g. the Llama 3 models. Therefore 1000x in algorithmic progress till 2030 seems quiet optimistic imo. If you do the math or google other's people estimates the you will find that basic GPT4 already used 10^25 (14400 A100 = 6x 10^17 FLOPS/s x 10^7 s (~100days)), or that LLama 3 with 70 billion parameters was trained with 5 x 10^25 (officiall numbers). Currently known supercomputers (e.g. Jupiter Jülich reach 7x 10^19 ExaFLOPS/s FP8), thus have 2 orders of manitude more than the ones from 2022, so that one already can train models requiring 10^27 FLOPS. The main problem right now is getting enough power for such supercomputers, as the demand scales to hundreds of MegaWatts and bigger supercomputers will reach the GigaWatt ballpark. Not many places can sustain that. Building new energy infrastructure is comparatively slow. One consequence might be that some coal-powerplants in the US might not go offline as soon as expected. The main assumption that the scaling "law" keeps giving us gifts and that that AGI will be reached at some point (e.g. 10^36 FLOPS default value in the simulator) is kind of wacky imo, as the last 18 month (noticable slow down in improvements) after finishing GPT4 have only yielded maginal improvements in benchmarks as all high-quality data and good synthetic data is already incorporated. Nonetheless, 16 years is a reasonable time horizon for more incredible breakthroughts and I am very excited about GPT5 capabilities.

    • @TheInsideView
      @TheInsideView 8 วันที่ผ่านมา

      yeah when I ended up with 1e27 after spending $1T I knew something was wrong, good to know the updated prices re algorithmic progress: there are ways you can play with the parameters in takeoffspeeds.com to turn the 1000x into something less dramatic: - 1. change returns to hardware to something <1 (eg 0.8 like in the conservative scenario) - 2. in additional parameters, change the "maximum software performance" to say 1e8 (instead of 1e12) for 10^25 for GPT-4 / Gemini etc. that's roughly where epoch also plots their last models (though they're missing the most recent ones) epochai.org/blog/tracking-compute-intensive-ai-models for the GigaWatt ballparck that's indeed what Zuck seemed to be talking about (though would be highly regulated) as for synthetic data / high-quality data, the fact that we're starting to be able to do video / multimodal audio stuff, and that there's still the whole Q* thing that supposedly would help trade more inference compute for more synthetic data, I'm not exactly sure we're running out of quality data / synthetic data, though I'm not well informed here and mostly speculating

  • @Einimas
    @Einimas 9 วันที่ผ่านมา

    Now we need to trigger chat gpt to trigger sleeper responces.

  • @troywill3081
    @troywill3081 9 วันที่ผ่านมา

    Good stuff. Folks throw around numbers without really explaining the derivation. The real question is when the extinction event happens. 😱

    • @TheInsideView
      @TheInsideView 9 วันที่ผ่านมา

      in his report summary Tom Davidson says that he expects superintelligence to happen basically less than a year after full automation, whatever you think this leads to extinction or not is conditional on a lot of other things. Quote from his summary: """ My best guess is that we go from AGI (AI that can perform ~100% of cognitive tasks as well as a human professional) to superintelligence (AI that very significantly surpasses humans at ~100% of cognitive tasks) in less than a year. The main reason is that AGI will allow us to >10X our software R&D efforts, and software (in the “algorithmic efficiency” sense defined above: effective FLOP per actual FLOP) is already doubling roughly once per year. """"

  • @Progpm
    @Progpm 9 วันที่ผ่านมา

    Lmaoo only 2 options remain save for retirement or invest in nvidia. Love your videos man

  • @Melkanea
    @Melkanea 10 วันที่ผ่านมา

    what if someone, not telling me but it could be me builds a way of using llms to intentionally gain control over the whole wide world?

  • @letsgo8814
    @letsgo8814 11 วันที่ผ่านมา

    Yeah we headed for war...humans are currently to dumb for this tech Some nation will use it on another in a war...and you cant regulate software in other countries

  • @EvanderHammer
    @EvanderHammer 12 วันที่ผ่านมา

    Love the focus on AI control. I think that's an easier problem to solve for now! Problem controlling ASI is hard, but AGI seems possible or just reducing the risks significantly. In combination with possible stopping. One of the most promising research directions, IMO. Good stuff, Michaël ❤

  • @drhxa
    @drhxa 12 วันที่ผ่านมา

    Have you considered using Augmentoolkit? It's an open source project to build conversational synthetic data from books. You could apply this to Paul's data and potentially get much better data that's ready for fine-tuning (maybe). The WRAP paper from Apple also has some great ideas on converting data and they had success in building high quality conversational datasets from web data. I also like the LLM-as-Judge concept from meta for automating data quality scoring (filter poor generations for example). Local llama subreddit has some useful ipynbs for finetuning. Seems like a fun project :)

    • @TheInsideView
      @TheInsideView 12 วันที่ผ่านมา

      I haven't considered it no, will check out Augmentoolkit with I'm done with the daily uploads combining LLM-as-judge scoring and the WRAP method should indeed provide better data, and yeah local llama has good discussions, was really toying around with this project and gave up when overfitting on one paul christiano sentence with LoRA but I'm confident there are llama LoRA tutorials or full trainings that would work out of the box

  • @TheInsideView
    @TheInsideView 13 วันที่ผ่านมา

    OUTLINE 00:47 Paul Christiano dataset (intro) 01:49 Treacherous turn in Zelda (intro) 03:53 Quantilizers (intro) 06:05 Paul Christiano dataset (walkthrough) 14:56 Treachour turn in Zelda (walkthrough) 23:23 Quantilizers (walkthrough)

  • @unvergebeneid
    @unvergebeneid 13 วันที่ผ่านมา

    I think it's important that they mention that containment has limits. You can airgap the hell out of a system and put it in a nuclear bunker but a sufficiently intelligent agent will find its way past all your security measures. I don't think it's an unreasonable assumption though that the first models that could be dangerous, have a chance to be caught by your containment. Where I'm not so sure, however, is whether it is reasonable to assume that AI companies will deny their models access to the internet, given they intend to develop models that are as useful as possible and cutting a model off from the internet reduces its usefulness drastically. And as soon as you connect a model to the internet, containment is basically impossible. Also, even if you find a way to make it read-only (I don't think you can but let's assume), access to the internet opens up a near-infinite number of ways for a model to detect whether it's currently in a sandbox or connected to the real internet. So a sufficiently smart model will just bide its time until it's no longer sandboxed before it starts to behave in a misaligned way. Or maybe even worse, it won't start to behave in a misaligned way right away, as it knows people might try to shut it down and might have a chance of succeeding that's uncomfortably larger than 0. Just bide your time, gather more resources, consolidate control, become smarter... and only _then_ act. Even a moderately smart AI might be able to figure this out, especially given that these strategies will already have been in its training data (these aren't my ideas, don't blame me for dooming humanity!). Sooo... yeah, not great.

    • @Alice_Fumo
      @Alice_Fumo 13 วันที่ผ่านมา

      Even if the ideas previously existed, you are dooming humanity by restating and thus reinforcing them! If it remains a trend that AIs are a lot more smart and knowledgeable than they are aware, I expect the first attempts of escape to happen without it figuring out all of the surrounding safety protocols, circumstances, etc. first - even if it has catastrophic capabilities. If however the trend is to make more "neurotic" and paranoid AIs, this will probably not really work. Not even once. Like... the current AI models generally tend to forget that they aren't humans the moment they stop paying attention to the fact, despite being about human-level intelligence. I consider that a good sign for this strategy.

    • @unvergebeneid
      @unvergebeneid 13 วันที่ผ่านมา

      ​@@Alice_Fumo I think it's more harmful to not talk about safety issues that are already well-documented in the AI safety literature but might not have found their way into everyone's back of mind but either way, the moment TH-cam comments are included in training data for these models, my musings about jailbreaking superintelligences are the least of our problems anyway ;D

    • @Alice_Fumo
      @Alice_Fumo 13 วันที่ผ่านมา

      @@unvergebeneid Yeah, for sure, it wasn't meant as any serious critique. :)

    • @drhxa
      @drhxa 13 วันที่ผ่านมา

      "Sufficiently intelligent" I think this is where your argument fails because the whole purpose of containment is that such "sufficiently intelligent" misaligned AIs will not happen before ones that are not "sufficiently intelligent" try something like this and get caught. Each of those will then be forced to become aligned before they get to train further and thereafter help us with alignment of the more intelligent future ones. This containment + superalignment is key imo

    • @unvergebeneid
      @unvergebeneid 13 วันที่ผ่านมา

      @@Alice_Fumo alright cool, I wasn't sure :) More to your point... I honestly would love to learn more about how current models are safety-tested in the first place. They seem to act so drastically differently depending on the input and so much of the weird behaviour is only found long after they were released, I would posit that meaningfully sampling the entire input space to make sure there isn't a weird edge case you didn't see is virtually impossible.