Llama-3.3: The BEST Opensource LLM EVER! Beats GPT-4o! (Fully Tested)

แชร์
ฝัง
  • เผยแพร่เมื่อ 17 ธ.ค. 2024

ความคิดเห็น • 48

  • @intheworldofai
    @intheworldofai  10 วันที่ผ่านมา +2

    Want to HIRE us to implement AI into your Business or Workflow? Fill out this work form: td730kenue7.typeform.com/to/WndMD5l7
    💗 Thank you so much for watching guys! I would highly appreciate it if you subscribe (turn on notifcation bell), like, and comment what else you want to see!
    📆 Book a 1-On-1 Consulting Call WIth Me: calendly.com/worldzofai/ai-consulting-call-1
    🔥 Become a Patron (Private Discord): patreon.com/WorldofAi
    🧠 Follow me on Twitter: twitter.com/intheworldofai
    Love y'all and have an amazing day fellas. Thank you so much guys! Love yall!

  • @lancemarchetti8673
    @lancemarchetti8673 10 วันที่ผ่านมา +8

    I used 3.3 in glhf last night to build a qrcode storage concept..After 4 hours I was so excited because the app actually worked.
    3.3 was particularly good at figuring out why my qrcode outputs from the design grid were not passing qrcode scanning tests. It tuned out that there was a faint grey 1px outline bordering each cell that was exported to the respective image format. I've noticed that keeping explanations simple in the prompting is key.
    Also, just paste issues from the console without explaining. The llm is pretty sharp at figuring out the issue.
    Overall...I'm impressed.

    • @AaronBlox-h2t
      @AaronBlox-h2t 9 วันที่ผ่านมา

      a qrcode storage concept?

  • @jackinsights
    @jackinsights 10 วันที่ผ่านมา +4

    Great analysis mate.

  • @MatthewSanders-l7k
    @MatthewSanders-l7k 10 วันที่ผ่านมา

    Love the enthusiasm for Llama 3.3! I'm definitely trying it out, cheaper and better than GP4 Omni is a win-win for me. Huge props to Meta AI team!

  • @pr0d1gyvisions74
    @pr0d1gyvisions74 10 วันที่ผ่านมา +3

    That butterfly looked like Mr Hankey with wings. 🤣🤣

    • @intheworldofai
      @intheworldofai  10 วันที่ผ่านมา +1

      HAHAHA south park ref

  • @jeffwads
    @jeffwads 10 วันที่ผ่านมา +1

    LM Studio has their own version of the model. Tested the 8bit and it is fantastic. Refactored a large 1200 line class into 650 lines including notes. Only a few minor adjustments and it worked beautifully.

    • @SahilP2648
      @SahilP2648 10 วันที่ผ่านมา

      Squashing lines of codes doesn't mean it's more performant or even cleaner. I am not sure why you mentioned only lines of code and nothing else.

  • @AlexJohnson-g4n
    @AlexJohnson-g4n 10 วันที่ผ่านมา

    Llama 3.3 sounds like a beast! Can't wait to try it out and see how it compares to GP4 Omni. Props to Meta AI team for making AI more accessible

  • @julienarpin5745
    @julienarpin5745 10 วันที่ผ่านมา +11

    It performs no better than Llama 3.1 on my private benchmark of verbal intelligence based on years of items encountered in the IQ testing community. My benchmark has matched prevailing benchmark leaderboards for years but is now diverging, indicating that these LLMs are gaining skillsets but not true intelligence, which is the ability to perform work on arbitrarily distant symbols in concept space, relations modelling, or sensitivity to analogies.

    • @marcus-b4x3h
      @marcus-b4x3h 10 วันที่ผ่านมา +1

      Make a video ❤

    • @pythonate2906
      @pythonate2906 9 วันที่ผ่านมา +2

      they are no more than freaking parrots
      if something is in their training set, you get a reasonable answer
      otherwise it's either BS either FOS, depends how deep it hallucinates
      no creativity whatsoever, just parroting its training set
      like a 5 yo with a huge SSD

    • @julienarpin5745
      @julienarpin5745 9 วันที่ผ่านมา

      @pythonate2906 Not exactly true since cranking the temperature up changes some function that enables novelty, which is arguably creativity as a whole. It's all about how you traverse the latent space with prompting and fine-tuning

    • @threepe0
      @threepe0 6 วันที่ผ่านมา

      @@pythonate2906you are no more than a parrot. How you use agents and teams is limited only by your imagination.

    • @pythonate2906
      @pythonate2906 6 วันที่ผ่านมา

      ​@@threepe0I guess stating the truth is disturbing. Not just me saying these things. But I wouldn't waste my time arguing.
      If something is NOT in the training set, parrots won't come up with it.
      I am testing NEW things, never been tested before, unlike your trivial daily BS.

  • @intheworldofai
    @intheworldofai  6 วันที่ผ่านมา

    Gemini 2.0 Flash: BEST LLM Ever! Beats Claude 3.5 Sonnet + o1! (Fully Tested): th-cam.com/video/lED0yLrUelM/w-d-xo.html

  • @intheworldofai
    @intheworldofai  10 วันที่ผ่านมา +1

    [Must Watch]:
    Qwen-2.5: The BEST Opensource LLM EVER! (Beats Llama 3.1-405B + On Par With GPT-4o): th-cam.com/video/yd0kgDwkfz0/w-d-xo.htmlsi=Uh2eCpIWYpcY54Hq
    DeepSeek-v2.5: BEST Opensource LLM! (Beats Claude, GPT-4o, & Gemini) - Full Test: th-cam.com/video/mvpkZ1yFy7o/w-d-xo.htmlsi=NR9ChO50-HKJW9Cb
    Bolt.New + Ollama: AI Coding Agent BEATS v0, Cursor, Bolt.New, & Cline! - 100% Local + FREE!: th-cam.com/video/ZooojV4ZDMw/w-d-xo.html

  • @HansKonrad-ln1cg
    @HansKonrad-ln1cg 6 วันที่ผ่านมา

    i have a 1 gallon jug, a 10 gallon jug, a 170 gallon jug and a 200 gallon jug. i want to measure exactly 173 gallons. how do i do that?
    all the models i have tested have failed this. the solution is obviously to pour a full 170 gallon jug into the 200 gallon one and then add a full 1 gallon jug three times to that. all the models either overthink it hopelessly or dont even add the right amount or want to fill jugs more than they can hold, etc. pretty sobering experience in case you got too excited about agi and stuff.

  • @intheworldofai
    @intheworldofai  8 วันที่ผ่านมา

    Warp: AI Terminal - FREE AI Coding Agent BEATS Cursor, V0 & Bolt! (Generate Full-Stack Apps!): th-cam.com/video/uRquE0FjvFk/w-d-xo.html

  • @ABUNDANCEandBEYONDATHLETE
    @ABUNDANCEandBEYONDATHLETE 10 วันที่ผ่านมา

    Idk if i downloaded a bad version but my 4090 liquid, 3970x w/256GB RAM is only pumping out like 0.5-2 tokens a sec. Thats not functional to even benchmark if its too slow.
    The qwen2.5 32B coder is like 30-60 tokens a sec.

  • @alinayossimouse
    @alinayossimouse 8 วันที่ผ่านมา

    how is that butterfly a pass? Maybe you should ask Llama 3.3 how many wings a butterfly has

  • @UCs6ktlulE5BEeb3vBBOu6DQ
    @UCs6ktlulE5BEeb3vBBOu6DQ 10 วันที่ผ่านมา +2

    QWEN2.5 72B and QwQ both ace the train question (tried it while watching this)

    • @intheworldofai
      @intheworldofai  10 วันที่ผ่านมา +1

      QwQ and the new deepseek reasoning model is quite impressive. Great alternative to o1

    • @UCs6ktlulE5BEeb3vBBOu6DQ
      @UCs6ktlulE5BEeb3vBBOu6DQ 10 วันที่ผ่านมา +1

      @@intheworldofai QwQ is better at coding, 72B is better at creating the code. If one could, he'd get 72B to create the program structure and tell QwQ to code it.

    • @ABUNDANCEandBEYONDATHLETE
      @ABUNDANCEandBEYONDATHLETE 10 วันที่ผ่านมา

      What lm-studio settings for this model should I use? The Q4_K_S is so slow for my machine.
      Loading 131072 context is laughingly slow. 4090 with 256GB and 3970X 32core AMD. Other models have no issues.

    • @intheworldofai
      @intheworldofai  10 วันที่ผ่านมา

      @@ABUNDANCEandBEYONDATHLETE Try decreasing the batch size to a smaller valuery setting. Plus you can also set model parallelism to False

    • @UCs6ktlulE5BEeb3vBBOu6DQ
      @UCs6ktlulE5BEeb3vBBOu6DQ 10 วันที่ผ่านมา +2

      @@ABUNDANCEandBEYONDATHLETE you are running a model that is slow on gpu with vram speed under 600gb/s on a system that offload too much load on memory that operate under 200gb/s. You are paying the poverty tax with your time. 72B at Q5 runs at 5-6 token/s for me. Its slow but at least I do not offload anything. Get yourself some other gpus and few PCI-e risers and increase your total vram. That system of yours is overkill for a server but its almost useless for LLM because of memory bandwith. Sorry

  • @lmaoanatorhilarious2582
    @lmaoanatorhilarious2582 10 วันที่ผ่านมา +2

    Does this have reasoning

    • @intheworldofai
      @intheworldofai  10 วันที่ผ่านมา +3

      Yup. Logical, analogical, and common sense reasoning

    • @tomgreen8246
      @tomgreen8246 10 วันที่ผ่านมา +2

      Sort of. Not in the o1 Preview sense... let alone o1 Pro (oh. my. God). Yes I dropped the money for it.

    • @intheworldofai
      @intheworldofai  10 วันที่ผ่านมา +3

      @@tomgreen8246 wow let me know how the tests go fro the o1 pro!

    • @lmaoanatorhilarious2582
      @lmaoanatorhilarious2582 9 วันที่ผ่านมา

      @@tomgreen8246 yeah I want that o1 reasoning. How do you like o1 pro? How does it differentiate from o1 preview

  • @HikaruAkitsuki
    @HikaruAkitsuki 6 วันที่ผ่านมา

    They basically downsized 405B to 72B?

    • @intheworldofai
      @intheworldofai  6 วันที่ผ่านมา

      Yes, it's a smaller model, but it's still incredibly powerful and efficient.

  • @TheVisionaryX
    @TheVisionaryX 9 วันที่ผ่านมา

  • @relexelumna5360
    @relexelumna5360 5 วันที่ผ่านมา

    Knowledge limited to Dec 2023 🥴

  • @prashantbhardwaj6322
    @prashantbhardwaj6322 10 วันที่ผ่านมา +1

    GPT 4o-MINI is $0.15/M tokens, how is this 25 times cheaper?

    • @intheworldofai
      @intheworldofai  10 วันที่ผ่านมา +3

      I said gpt-4o. I never mentioned gpt-4o mini

    • @prashantbhardwaj6322
      @prashantbhardwaj6322 10 วันที่ผ่านมา

      @@intheworldofai at 00:30 mark, maybe I heard it wrong.

    • @intheworldofai
      @intheworldofai  10 วันที่ผ่านมา +3

      @@prashantbhardwaj6322sorry brotha I think you heard wrong.
      I said “25x cheaper than gpt-4o” not gpt-4o mini. There two different models.