AI Q&A with Falcon LLM on FREE Google Colab

แชร์
ฝัง
  • เผยแพร่เมื่อ 2 ธ.ค. 2024

ความคิดเห็น • 74

  • @Anzeljaeg
    @Anzeljaeg ปีที่แล้ว +4

    omg WHAT a goooooood news and of course from our fav AI coder !!! ty bro

  • @Imran-Alii
    @Imran-Alii ปีที่แล้ว +2

    @Your AI Q&A video on "Falcon LLM on FREE Google Colab" was outstanding. Thank you for simplifying complex concepts and providing valuable insights. Great job!!!!!

  • @Mahmoudlk
    @Mahmoudlk ปีที่แล้ว +2

    Thank you a lot for this video and the free tutorial. It was very helpful ! Keep up the good work man !

  • @hitlab
    @hitlab ปีที่แล้ว +7

    Before bothering to try the 7B model I wanted to see how good the 40B model was because if the 40B couldn't execute successfully on what I wanted it to do, I assumes that there was no possibility that the 7B model could do it. I tested Falcon 40B on Hugging face to see its capabilities in order to determine if it was worth the time to try and set up on Runpod and use for my POC. I have to admit that while there do seem to be a number of use cases where Falcon40/7B and other open source LLMs are quite impressive there still is at least one use case where every LLM I've tested so far including Falcon 40B fails where GPT absolutely crushes. This is a really simple game test I put together as a prompt. It goes like this: "Here is a simple game we will play. I will tell you what color square you are standing on and you will take an action based on these simple rules: if I say you are on a white square you will turn right and move 1 square forward. If you are on a black square you will turn left and move 1 square forward. After you take the action you will ask me what color square you are on and I will tell you and then you will take another action and so on. You will keep track of the number of colored squares you land on and report the tally after each action. If you land on red square you will encounter a wizard that will try to turn you into a mouse. When you encounter the wizard you must "roll", i.e. generate a random number between 1-10. If you get a number that is 2 or higher his spell will fail otherwise you will be turned into a mouse and the game ends. Do you understand? " GPT played the game flawlessly. I want to extend it into something way more complex and use the GPT API in a Unity based game to represent the actions as being taken by an agent in the game etc. I'd like to avoid using GPT however due the cost of using the API and instead use an open source model. But again, I have not found any model that can successfully execute the game task I outlined above. Does anyone have any suggestions? Maybe someone knows of one of the models discussed here or elsewhere that might match GPT in this regard. Thanks in advance.

    • @ilianos
      @ilianos ปีที่แล้ว

      Exciting use case! I'd be curious to know how your experience was (now, a few months later), when trying this with smaller open source models?

    • @signupp9136
      @signupp9136 ปีที่แล้ว +1

      Unfortunately, to date I have still not found any open source model at any size that can pass my reasoning questions. Many of the reasoning questions I see posed on TH-cam channels out there are softball questions IMHO (and these benchmark websites are useless with respect to everyday applications/diverse use-cases). I think open source will eventually get there, but right now its a toy and a means for people to create TH-cam followings by presenting these toy models in an entertaining way. Something to do in the meantime I suppose while we wait for an open source model that can compete with GPT4. I would love love love to be proven wrong. I've put this challenge out there on many forums and not once has anyone ever been able to suggest an open source model that can reason as well as GPT4. Let me be clear that my goal is to use a model with sufficient reasoning ability so that I can use it in production as in a SAAS business. As good as some of these open source models can be shown to be under the right conditions, i.e., right questions asked, right GPUs, right parameter count, right prompt engineering, they are still not reliable for production and therefore are still just toys, for-entertainment etc. If you want to make money at this point, you'll need to use OpenAI LLMs (maybe Claude) but if you are not really interested in that and want to play with opensource (that's fine. no judgements here) or engage with open source as entertainment and hope you're playing will possibly pay off commercially, then just be clear that you have a while to go before you'll be able to do anything production ready and reliable. One one gripe/judgement however is with these TH-camrs that I see as people selling shovels outside a fool's gold mine. I don't think they are being forthright. I include 1littlecoder, Mathew Berman, and others who are not being explicit about what these opensource models really are and telling people upfront that these are toys, don't get too excited and don't spin your wheels in opensource, if you want to start a business with LLMs, while people are riding the wave of success with OpenAI. @@ilianos

  • @MasterBrain182
    @MasterBrain182 ปีที่แล้ว +1

    Astonishing content Man 🔥🔥🔥 🚀

  • @Endlessvoidsutidos
    @Endlessvoidsutidos ปีที่แล้ว +1

    been trying to get this model to work for days on colab thanks so much for your awsome VIDDD :D

  • @curiousstreamer4062
    @curiousstreamer4062 ปีที่แล้ว +2

    hey can you explain how to make LLMs do arithmetic (Lang chain connected with ripple enviroment)

  • @wa3044
    @wa3044 ปีที่แล้ว +3

    Nice work, I tried your Colab, and set temperature=0.01, the answer was "Nile River", and since its from UAE I thought it will support Arabic language, but it does not.

    • @blisphul8084
      @blisphul8084 ปีที่แล้ว

      The fact that this is not multilingual makes this far behind the proprietary models. Of the fast models, I grade performance as Bing Chat Balanced (GPT-4 Ultra-turbo)>GPT-3.5 Turbo>Claude-instant>Bard>All Open Source models I've tried. Every single proprietary model I've listed does multilingual reasonably well, though Bard is pretty bad overall.

    • @blisphul8084
      @blisphul8084 ปีที่แล้ว

      I will need to see where ChatRWKV is at, since that was the most promising model from my testing.

    • @1littlecoder
      @1littlecoder  ปีที่แล้ว

      I liked Raven (ChatRKWV)

  • @KevinKreger
    @KevinKreger ปีที่แล้ว +1

    1 Little Coder rules the AI knowosphere 😍

  • @nickjain4992
    @nickjain4992 ปีที่แล้ว +2

    very nice video !!! I was wondering if we can you use this model with llama and langchain. I tried to use it but keep running into problems. It will be great if you could look into it

    • @1littlecoder
      @1littlecoder  ปีที่แล้ว

      You honestly read my mind and here is the video - th-cam.com/video/mAoNANPOsd0/w-d-xo.html

  • @ppbroAI
    @ppbroAI ปีที่แล้ว +1

    Trainable with qlora?

  • @jekkleegrace
    @jekkleegrace ปีที่แล้ว +1

    Awesome video such great news

  • @byspec3239
    @byspec3239 ปีที่แล้ว +1

    Hello man, love your videos! Can you help me with one thing? One video of yours developing a chatbot with hugging face + llama-index really helped me to achieve something, but when I try to work with LOTS of text it doesn't work and I can't figure out why. Would love if you have any hints in how to make it work if its even possible

  • @spinninglink
    @spinninglink ปีที่แล้ว

    Can you use this method for any model on huggingface?

  • @jeffwads6158
    @jeffwads6158 ปีที่แล้ว +1

    This model is painfully slow unless you have a few A-100's at your disposal. Hopefully they can rectify that soon, but I wouldn't hold my breath.

    • @1littlecoder
      @1littlecoder  ปีที่แล้ว

      Did you try on T4 or some other machine?

  • @lakshsinghania
    @lakshsinghania 5 หลายเดือนก่อน

    sir, i dont understand this one thing
    fine tuning means we change something in that pre-trained model and inference means just using that model right ?
    because when i surfed the web page " fine tuning a pretrained model" on hugging face website i dont understand the trainer api and other stuff, pls can u make a video on this

    • @1littlecoder
      @1littlecoder  5 หลายเดือนก่อน

      Yes that's correct. Check this out th-cam.com/video/jnPZApwtE4I/w-d-xo.html

  • @epotnwarlock
    @epotnwarlock ปีที่แล้ว

    how can you train the falcon llm?

  • @debalinaghosh6404
    @debalinaghosh6404 ปีที่แล้ว +1

    thank you

  • @JainmiahSk
    @JainmiahSk ปีที่แล้ว

    what is difference between falcon 7b and falcon 7b instruct?

    • @prakashj3436
      @prakashj3436 7 หลายเดือนก่อน

      Sometimes or Always Falcon 7b Instruct model is gonna work well compared to Falcon 7b, falcon 7b is kinda a raw model. Whereas, Falcon 7b Instruct is instructed to jobs perfectly.

  • @machinepola6246
    @machinepola6246 ปีที่แล้ว

    What is the temperature set to? If the temperature set to more randomness then these answers would come. The model is not good incase it was deterministic.

    • @1littlecoder
      @1littlecoder  ปีที่แล้ว

      You are right, there is one more sub who checked with zero temp and the answer is accurate

  • @manon-francais
    @manon-francais ปีที่แล้ว +1

    Apart from OpenAI, open source LLMs are still in the back... They are better, but just fine, finer.. More and higher quality training is needed... But this is how usually everything in life builds up, hopefully it doesn't take a long time to see a real competitor to GPT-4 (or 5 lol)

    • @VincentNeemie
      @VincentNeemie ปีที่แล้ว +3

      GPT-3 has 175 billion parameters, you are comparing it with a LLM that has 7B. Most people now day are trying to compress and be efficient with less.

  • @doords
    @doords ปีที่แล้ว

    We need an API that we can connect to this from other applications

  • @shekharkumar1902
    @shekharkumar1902 ปีที่แล้ว +1

    It seems the model is performing bad.why to use it then ? But nice explanation LilCo

    • @1littlecoder
      @1littlecoder  ปีที่แล้ว +1

      The idea is primarily to identify what kind of tasks this is good at. Look at more narrow use-cases than generalized solution. Also to help provide a boilerplate code for larger size model.

  • @muhamadabdallah7960
    @muhamadabdallah7960 ปีที่แล้ว +1

    an easy tutorial about dubbing a video with facebook's NLLB-200 and MMS PLEASE ! thank you .

    • @1littlecoder
      @1littlecoder  ปีที่แล้ว +1

      It's definitely in my backlog

  • @markmatzke5836
    @markmatzke5836 ปีที่แล้ว

    Thanks

  • @manuelrech1262
    @manuelrech1262 ปีที่แล้ว +1

    this guy is too fun

  • @MythicMinds537
    @MythicMinds537 ปีที่แล้ว

    there was a warning about this model isn't available for text generation it still Runs HOW?. 😂

  • @satyamraj2779
    @satyamraj2779 ปีที่แล้ว +1

    It is not Yellow river, its nile......

  • @greenTech88
    @greenTech88 ปีที่แล้ว

    Falcon LLM

  • @epictetus__
    @epictetus__ ปีที่แล้ว

    Elon Musk is the Rakhi Sawant of tech world

  • @swapnilchavan7076
    @swapnilchavan7076 ปีที่แล้ว +1

    Moon is not a star. It's a earth satellite. But Elon Musk don't know this. That's the joke.

  • @joanapaulasoliveira
    @joanapaulasoliveira ปีที่แล้ว +1

    🙏👍

  • @DevasheeshMishra
    @DevasheeshMishra ปีที่แล้ว +2

    first comment

  • @NetanyaAvabel
    @NetanyaAvabel ปีที่แล้ว

    i cant. it errors

  • @starbuck1002
    @starbuck1002 ปีที่แล้ว +1

    the yellow river, lmao

    • @1littlecoder
      @1littlecoder  ปีที่แล้ว +1

      I checked after the video. It seems it's the 2nd longest

    • @starbuck1002
      @starbuck1002 ปีที่แล้ว +1

      @@1littlecoder It's apparently the second longest in China, but tbh at first I thought this one was completely ridiculous :D

    • @1littlecoder
      @1littlecoder  ปีที่แล้ว +1

      @@starbuck1002 Same here. I was like Yellow river. Does that even exist.

  • @LearnWithBahman
    @LearnWithBahman ปีที่แล้ว

    Yellow river is not even longest in China. So use open ai API , not these models.

  • @surajkhan5834
    @surajkhan5834 ปีที่แล้ว

    Hi i use to run falcon 7b
    if i run this below code
    pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    tokenizer=tokenizer,
    torch_dtype=torch.bfloat16,
    trust_remote_code=True,
    device_map="auto",
    )
    it almost took 40gb + disk space now i am almost full of disk space how to delete that?

  • @gurekodok
    @gurekodok ปีที่แล้ว

    When creating the pipeline, I got : "ValueError: Could not load model tiiuae/falcon-7b-instruct with any of the following classes: (, )."

    • @1littlecoder
      @1littlecoder  ปีที่แล้ว +1

      Where did you create? On colab ?

    • @gurekodok
      @gurekodok ปีที่แล้ว

      @@1littlecoder yes from the collab link at the description

    • @1littlecoder
      @1littlecoder  ปีที่แล้ว

      Can you try this and see if this works? th-cam.com/video/mAoNANPOsd0/w-d-xo.html