Anthropic’s new 100K context window model is insane!

แชร์
ฝัง
  • เผยแพร่เมื่อ 8 ก.พ. 2025
  • Anthropic released a new LLM with a 100K token context window. In this video I'll explain what this means and we look at a demo.
    Learn more about Anthropic's new model: www.anthropic....
    Get your Free Token for AssemblyAI👇
    www.assemblyai...
    ▬▬▬▬▬▬▬▬▬▬▬▬ CONNECT ▬▬▬▬▬▬▬▬▬▬▬▬
    🖥️ Website: www.assemblyai...
    🐦 Twitter: / assemblyai
    🦾 Discord: / discord
    ▶️ Subscribe: www.youtube.co...
    🔥 We're hiring! Check our open roles: www.assemblyai...
    ▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
    #MachineLearning #DeepLearning

ความคิดเห็น • 64

  • @pmarreck
    @pmarreck ปีที่แล้ว +23

    You literally picked THE PERFECT podcast to summarize. That's literally the one I've been trying to find a time window to digest (having a 2 year old makes this extremely difficult!)
    Ausgezeichnet!!

    • @AssemblyAI
      @AssemblyAI  ปีที่แล้ว +1

      haha glad to hear that!

  • @julian-fricker
    @julian-fricker ปีที่แล้ว +19

    Wow, spent weeks learning about embeddings and this throws it all out the window. 😂

    • @absta1995
      @absta1995 ปีที่แล้ว +8

      Embeddings will always be useful tbf

    • @lori208
      @lori208 ปีที่แล้ว +5

      @@absta1995 I totally agree with you, Actually, embeddings can be used on their own for Q&A systems. Not as chatty as LLMs but much, much cheaper, as ingesting 100k tokens in a single prompt will surely be costly.

    • @AssemblyAI
      @AssemblyAI  ปีที่แล้ว

      haha

    • @seanfeng5448
      @seanfeng5448 ปีที่แล้ว

      Still useful for over 100k tokens.. for now 😁

    • @JorgeZentrik
      @JorgeZentrik ปีที่แล้ว +3

      No, no. Throw it *in* the window
      In the context window 😊

  • @luis96xd
    @luis96xd ปีที่แล้ว +4

    Great video, this model looks so powerful! Thanks for sharing!

  • @6little6fang6
    @6little6fang6 ปีที่แล้ว +2

    Amazing, what a time to be alive!

  • @l0ltaha
    @l0ltaha ปีที่แล้ว +2

    Great walkthrough/explanation and video format as usual!

  • @answerai
    @answerai ปีที่แล้ว +2

    Pretty incredible the speed of progress here

  • @laurentburdin9932
    @laurentburdin9932 ปีที่แล้ว

    Great demo - thx!! Best from Hamburg

  • @JazevoAudiosurf
    @JazevoAudiosurf ปีที่แล้ว +3

    how expensive is the claude-v1-100k model?

  • @time.5942
    @time.5942 ปีที่แล้ว +5

    So the advantage of Claude over e.g. OpenAI with LangChain are: 1) I don't have to summarize parts of larger documents to later retrieve the ones deemed relevant for answering a given question and thus 2) I don't have to worry about any kind of Vectorstore. Did I get this right?
    When using Claude, as the Input prompt contains the large text corpus, I'm feeding a lot more tokens during inference thus paying more?

    • @AssemblyAI
      @AssemblyAI  ปีที่แล้ว +1

      sounds correct

    • @rafaelfigueroa2479
      @rafaelfigueroa2479 ปีที่แล้ว +2

      It's correct conceptually and if you want to use personally to summarize papers and books for example. But in practical terms calling the API recurrently with 100k tokens it's not feasible cost wise for production arquitetures.

    • @kaynkayn9870
      @kaynkayn9870 ปีที่แล้ว +1

      @@rafaelfigueroa2479 hopefully in the future tokens cost becomes dirtcheap that this would be no issue.

  • @mjxxz
    @mjxxz ปีที่แล้ว +2

    Hello, the video is very good. I would like to ask if claude's api has similar SYSTEM prompts as in gpt3.5?Or there's a strong system of cue weights to keep him from forgetting to give him an identity or a situation?

  • @guptafamily1981
    @guptafamily1981 ปีที่แล้ว

    Do i need to always send the whole data[text] , always when doing prompt engineer

  • @martinsherry
    @martinsherry ปีที่แล้ว +2

    How long before you create an agent that listens to your youtube videos and then responds to the commenter questions that were already answered in your video ?

  • @terogamer345
    @terogamer345 ปีที่แล้ว +5

    That section about putting 100K tokens into context was eye opening, that means you could put the original 3 Star Wars scripts and ask it to write a whole sequel??

    • @AssemblyAI
      @AssemblyAI  ปีที่แล้ว +3

      yes!

    • @computerex
      @computerex ปีที่แล้ว +1

      @@AssemblyAI No, because the output is still constrained to like 2k tokens.

    • @SamWatkinsAU
      @SamWatkinsAU ปีที่แล้ว +3

      @@computerex so run it several times, output is produced sequentially anyway as I understand. It might get a bit expensive but likely cheaper than George Lucas at least.

  • @alioraqsa
    @alioraqsa ปีที่แล้ว

    Man, this is crazy

  • @jackzhang8712
    @jackzhang8712 ปีที่แล้ว

    Does it work just in the chatting interface without calling API now?

  • @MichaelScharf
    @MichaelScharf ปีที่แล้ว

    What was the response time of the prompts you demoed?

  • @duudleDreamz
    @duudleDreamz ปีที่แล้ว +4

    Has Claude improved over the past month? Last time I tested it, GPT4 was far superior. Little point switching unless the underlying Claude system has also improved. I'd rather use embeddings with a superior GPT4 model, than 100k context in mediocre Claude model

  • @DeepakSingh-ji3zo
    @DeepakSingh-ji3zo ปีที่แล้ว +1

    HI @AssemblyAI Excellent video as always! I just had a question is Anthropic API free to use or do we need to buy a subscription for this model?

  • @aurelienb9109
    @aurelienb9109 ปีที่แล้ว +1

    Yes, but how much does it cost if for each prompt we have to send the 100,000 tokens of the document to the API ? It seems quite expensive to use ...

    • @baggebilla
      @baggebilla ปีที่แล้ว

      Starting from 1.60 to 33 usd/ million tokens

    • @aurelienb9109
      @aurelienb9109 ปีที่แล้ว +1

      @@baggebilla hopefully, price for LLMs use will go down as the technology advances and we will even have the ability to do it locally for free with performant models on one's laptop...

  • @Moyano__
    @Moyano__ ปีที่แล้ว

    How much money does it cost to feed it 100k tokens?

  • @yusufkemaldemir9393
    @yusufkemaldemir9393 ปีที่แล้ว +2

    Hi,
    1- Do I have to pay for the API key?
    2- Can this read with pdf files? Thanks

    • @pmarreck
      @pmarreck ปีที่แล้ว +2

      If it can't read PDF's directly, there are plenty of services out there that will, for example, convert PDF's to Markdown etc.

    • @yusufkemaldemir9393
      @yusufkemaldemir9393 ปีที่แล้ว

      @@pmarreck thanks for the reply. Do I have to pay for the API?

    • @billnjoroge7088
      @billnjoroge7088 ปีที่แล้ว +4

      @@yusufkemaldemir9393 of course lol, inference isn't cheap. they did say their pricing is similar to their standard pricing rn

  • @tangobayus
    @tangobayus ปีที่แล้ว

    I requested access, got accepted, and then nothing. No email, no web page. Nada. Fail. Not ready for prime time.

  • @DistortedV12
    @DistortedV12 ปีที่แล้ว +1

    Does this get better if you give the same context multiple times?

  • @mjxxz
    @mjxxz ปีที่แล้ว

    cool!

  • @irshviralvideo
    @irshviralvideo ปีที่แล้ว +1

    there is a research paper showing 2M context size is possible.

    • @AssemblyAI
      @AssemblyAI  ปีที่แล้ว +4

      This model is probably just the beginning

    • @Moyano__
      @Moyano__ ปีที่แล้ว +2

      Finally we'll get the ending for the GoT novels!

    • @newyorthtimes4496
      @newyorthtimes4496 ปีที่แล้ว

      the 2M context size is not the same as what's happening here. The RMT paper just showcases the ability to retrieve facts and do very simple reasoning tasks (uses 1 or 2 facts at most and even with just that the performance drops quite a bit), while claude is claiming to be able to properly synthesize the data throughout the whole 100k context windox

  • @drgutman
    @drgutman ปีที่แล้ว

    The link is dead?

    • @AssemblyAI
      @AssemblyAI  ปีที่แล้ว +1

      now it's live!

    • @drgutman
      @drgutman ปีที่แล้ว

      @@AssemblyAI - Thank you. It works.

  • @calabisan
    @calabisan ปีที่แล้ว

    sounds amazing, but sorry, neither url exists nor code works as is.

  • @WilliamDye-willdye
    @WilliamDye-willdye ปีที่แล้ว +1

    Please stop over-using the word "insane" when describing an AI.
    We may need it someday.

    • @AssemblyAI
      @AssemblyAI  ปีที่แล้ว +1

      ok maybe you're right ;)

  • @ko-Daegu
    @ko-Daegu ปีที่แล้ว

    still no one can actually use it thou

  • @M1ndV0yag3r
    @M1ndV0yag3r ปีที่แล้ว

    Is the code available for free @AssemblyAI?

    • @alissonryan
      @alissonryan ปีที่แล้ว +1

      I would love the code too. I use Claude a lot more than ChatGPT and have been happy with the results.