I used LLaMA 2 70B to rebuild GPT Banker...and its AMAZING (LLM RAG)

แชร์
ฝัง
  • เผยแพร่เมื่อ 16 ส.ค. 2023
  • 👨‍💻 Sign up for the Full Stack course and use TH-cam50 to get 50% off:
    www.coursesfromnick.com/bundl...
    🐍 Get the free Python course
    go.coursesfromnick.com/python
    Hopefully you enjoyed this video.
    💼 Find AWESOME ML Jobs: www.jobsfromnick.com
    🤖 Get the Code:github.com/nicknochnack/Llama...
    Learn how to use Llama 2 70B Chat for Retrieval Augmented Generation...for FINANCE! Albeit in a hella haphazard way, oh, we'll also build a streamlit app while we're at it.
    Oh, and don't forget to connect with me!
    LinkedIn: bit.ly/324Epgo
    Facebook: bit.ly/3mB1sZD
    GitHub: bit.ly/3mDJllD
    Patreon: bit.ly/2OCn3UW
    Join the Discussion on Discord: bit.ly/3dQiZsV
    Happy coding!
    Nick
  • วิทยาศาสตร์และเทคโนโลยี

ความคิดเห็น • 206

  • @moondevonyt
    @moondevonyt 11 หลายเดือนก่อน +42

    first off, respect for the hustle and the in-depth breakdown of integrating llama with other tools
    really shows how much work goes behind the scenes
    that said, not sure why everyone's so hyped about all these new models when sometimes simpler and older architectures can do the trick
    but hey, if it's all about pushing boundaries and experimenting, you're killing it bro!

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน +6

      Thanks a mill moondev!! Yah at this point I'm just pushing to see where it's going, I started fine tuning this for some custom use cases and it looks hyper promising though!

    • @xt3708
      @xt3708 11 หลายเดือนก่อน +2

      Thanks so much for the detailed videos @@NicholasRenotte Can you make a video on fine tuning?

    • @vyrsh0
      @vyrsh0 10 หลายเดือนก่อน +3

      can you name some of the old models? so I can look them up and learn about them?

    • @ZombiemanOhhellnaw
      @ZombiemanOhhellnaw 10 หลายเดือนก่อน

      @@vyrsh0 @moondevonyt yes, I would like to learn what older models do the trick as well!

  • @malice112
    @malice112 11 หลายเดือนก่อน +7

    NIcholas I love your videos and your way of making learning about ML/AI fun! In your next video can you please show us how to fine-tune a LLM model! Thanks for all the hard work you do on making these videos!

  • @yudhiesh1997
    @yudhiesh1997 11 หลายเดือนก่อน +118

    You can’t load Llama2-70b on a single A100 GPU. Using full precision(float-32) would require 70billion * 4 bytes = 280GB of GPU memory. If you load it using float-16 it would reduce by half to 140GB. It finally worked cause you loaded it in int-8 which only requires 70GB of memory while the A100 has 80GB of GPU memory. If you wanted to load it in full/half precision you would need multiple GPUs and also need to leverage tensor parallelism whereby you slice the tensors across multiple GPUs.

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน +12

      I'm not sure that was it, I successfully loaded in half precision over 2xA100-80GB (didn't show the loading in the vid). But when I went to generate this is what I came up against: github.com/huggingface/transformers/issues/24056. Solid calcs though!

    • @sluggy6074
      @sluggy6074 11 หลายเดือนก่อน +9

      That's nice. I'll just have to settle for my quantized 70b LLMs that run hot and fast on my 4090.
      I think I can live with this.

    • @agusavior_channel
      @agusavior_channel 11 หลายเดือนก่อน +1

      Use petals

    • @seanhuver4813
      @seanhuver4813 10 หลายเดือนก่อน

      It runs nice @ 4bit precision on A6000.

    • @bubbleboy821
      @bubbleboy821 9 หลายเดือนก่อน

      What you meant to say was that you can load LLama2-70b on a single A100 GPU, you just have to run it in int-8.

  • @MikeAirforce111
    @MikeAirforce111 11 หลายเดือนก่อน +5

    This video was great. You have created a format that is very entertaining to watch! 🙌 Subbed!

  • @splitpierre
    @splitpierre 5 หลายเดือนก่อน +2

    Yeah, nice work!
    I've been playing around RAG as well, I can relate to all roadblocks and pain points.
    I'm trying to squeeze as much as possible so I can have a decent RAG, without any fancy GPU, consumer grade hardware running everything local, it's been fun/painful

  • @richardbeare11
    @richardbeare11 9 หลายเดือนก่อน

    Love your videos Nicholas. Watching this with my morning coffee, a few chuckles, and a bunch of "ooohhh riiiiiight!"s. Your vid bridged a bunch of gaps in my knowledge.
    Gonna be implementing my own RAG now 😎👍

  • @princechijioke247
    @princechijioke247 11 หลายเดือนก่อน +9

    Always looking forward to your videos...
    I've an MSc. In AI, but I still learn from you 👏🏼

    • @MikeAirforce111
      @MikeAirforce111 11 หลายเดือนก่อน +12

      I have a PhD and I am here as well 🤷‍♂

    • @siestoelemento4027
      @siestoelemento4027 10 หลายเดือนก่อน

      I guess I'm on the right path then
      @@MikeAirforce111

  • @hongyiyilim6830
    @hongyiyilim6830 9 หลายเดือนก่อน

    Great content! helped me alot with building my own open source model RAG

  • @dacoda85
    @dacoda85 11 หลายเดือนก่อน +3

    Love this style of video. Fantastic content as always mate. You've given me some ideas to try out. Thanks :)

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน +1

      🙏🏽 thanks for checking it out!

  • @FunCodingwithRahul
    @FunCodingwithRahul 11 หลายเดือนก่อน +3

    Incredible stuff done... thank you Nich.

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Anytime!! Glad you liked it Rahul!

  • @sinasec
    @sinasec 8 หลายเดือนก่อน

    well done. one of the best and compact tutorial I ever had. Thanks for providing the source code

  • @ShaneZarechian
    @ShaneZarechian 6 หลายเดือนก่อน

    Taking the viewers along the development and debugging ride is a cool style

  • @ricowallaby
    @ricowallaby 10 หลายเดือนก่อน

    Hi, just found your channel and enjoying it, but I can't wait till we have real Open Source LLMs, anyway keep up the good work, cheers from Sydney.

  • @kallamamran
    @kallamamran 11 หลายเดือนก่อน

    I so wish I could do this. Maybe not specifically THIS, but things like this. I wish I understood the underlying principles for making something like this work, Great video!!!

    • @kallamamran
      @kallamamran 10 หลายเดือนก่อน

      @@jimmc448 Ha ha ha...

  • @ShahJahan_NNN
    @ShahJahan_NNN 11 หลายเดือนก่อน +2

    please make a video on ocr on past question papers that can extract questions, and extract keywords and analyse with 10 years papers, and predicts upcoming questions

  • @shipo234
    @shipo234 11 หลายเดือนก่อน +3

    Nick this is insanely good, thank you for the effort

  • @wayallen831
    @wayallen831 11 หลายเดือนก่อน +2

    Great tutorial! Can you also help do a tutorial on setting up runpod to host the application on it? Found that part to be a bit confusing and would love a more thorough walk thru. Thanks for all you do!

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Ya, might do something soon and add it to the free course on Courses From Nick. I'm saving infra style/setup videos for the Tech Fundamentals course.

  • @accelerated_photon2265
    @accelerated_photon2265 10 หลายเดือนก่อน

    Love your videos , would love to deploy a model but the 70B compute is way too much do you have any idea or do you know any website where I can check compute requirements for the 7B model ? Just got my meta access last week thanks again for the video

  • @juanpablopereira1479
    @juanpablopereira1479 11 หลายเดือนก่อน +6

    I think "Amazing" falls short, the amount of knowledge, the fact that your using cutting edge Open source model and all of that in a really funny and light tone. Keep up the good work! I have a question, do you think is much harder to deploy that app into google cloud run compared with runpod?

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Thanks so much Juan! I can't imagine it would be running on a VM instance with GPUs attached. Could also separate out the LLM bit and run that solely on a GPU then just run the app on a basic linux instance!

  • @autonomousreviews2521
    @autonomousreviews2521 11 หลายเดือนก่อน +1

    Great share! Thank you for your persistence and giving away your efforts :)

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Anytime! Gotta share where I can!

  • @andreyseas
    @andreyseas 11 หลายเดือนก่อน +1

    Sick production value and great content!

  • @daniamaya
    @daniamaya 10 หลายเดือนก่อน

    wow! This is top-tier content. Thank you!

  • @PritishMishra
    @PritishMishra 11 หลายเดือนก่อน +1

    Amazing editing and content, learnt a lot.

  • @BudgetMow
    @BudgetMow 8 หลายเดือนก่อน

    thank you for this tutorial , although i am facing a slight issue in parsing tables from pdfs , i managed to allow the parser to take in multiple documents , and it is answering in a quick time , only issue with if the question is related to data within a table or some times data spanning multiple lines it fails to retrieve that data

  • @chrisweeks8789
    @chrisweeks8789 11 หลายเดือนก่อน +3

    All facets of your work are incredible! Are the context limits of llama2 similar to that of OpenAI?

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Thanks a mil! Would depend on which models you're comparing!

  • @himanshuahujaofficial7813
    @himanshuahujaofficial7813 4 หลายเดือนก่อน

    Nick, thank you so much for the great content. I’m new to AI and want to build an LLM for my startup, but I’m not sure where to start. Can you recommend something?

  • @yashsrivastava4878
    @yashsrivastava4878 10 หลายเดือนก่อน

    hey can it be done on chainlit along with LMQL and Langflow added to it, output shows pdfs file as a reference and scores based on whether its retrieves factual data or makes up its own answer

  • @kevynkrancenblum5350
    @kevynkrancenblum5350 11 หลายเดือนก่อน +1

    2:40 😂
    Thanks Nic the video is awesome ! 🤘🏽🤘🏽🤘🏽

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      LOL, stoked you liked it Kev!!

  • @ba70816
    @ba70816 11 หลายเดือนก่อน +1

    Really great content, you might have the most effective style I’ve ever seen. Well done. I can’t remember which video I saw where you spoke about your hardware setup. It’s cloud based isn’t it?

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Thanks a mil! This particular instance is cloud based, yup! It's all runpod, I used a remote SSH client to use the env with VsCode. Old HW vid might have been this: th-cam.com/video/GH1RuKguO54/w-d-xo.html

    • @ba70816
      @ba70816 11 หลายเดือนก่อน

      Would you consider a video showing the setup process you use?

  • @projecttitanium-slowishdriver
    @projecttitanium-slowishdriver 11 หลายเดือนก่อน +1

    Huge thanks for you videos. Nowadays I code, demonstrate, and perhaps lead AI, ML, DL, and RL development in 1300 + worker engineering and consulting company
    I am combining technical analysis tools (fem, CFD, MBS…) with AI to generate new digital business cases

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Ooooh, sounda amazing!

    • @projecttitanium-slowishdriver
      @projecttitanium-slowishdriver 11 หลายเดือนก่อน

      @@NicholasRenotte It is a 13 workers' digital business development group:) But thanks again mate

  • @ciberola285
    @ciberola285 10 หลายเดือนก่อน

    Hi Nicolas, are you planing to make a video on training OWL-ViT model?

  • @vanshpundirv_p_r9796
    @vanshpundirv_p_r9796 11 หลายเดือนก่อน

    Hey, Can you tell me minimum vram, ram and space required to load and inference from the model?

  • @tenlancer
    @tenlancer 10 หลายเดือนก่อน

    what is the response time for each query? and which GPU did you use for this app?

  • @emanuelsanchez5245
    @emanuelsanchez5245 10 หลายเดือนก่อน

    Hi!
    What was the performancee of the method?
    How many tokens per second with that deployment?

  • @ml-techn
    @ml-techn 11 หลายเดือนก่อน

    Hi, thanks for the video. which gpu are using? I want to buy and build a dl machine to play with llm.

  • @mohamedkeddache4202
    @mohamedkeddache4202 10 หลายเดือนก่อน

    please help me 😓
    ( in your videos of licence plate tensorflow)
    i have this error when i copy the train command in cmd :
    ValueError: mutable default for field sgd is not allowed: use default_factory

  • @jennilthiyam1261
    @jennilthiyam1261 8 หลายเดือนก่อน

    How can we set up llama=-2 on the local system with memory. Not just one time question but interactive conversation like online chatGPT

  • @nimeshkumar8508
    @nimeshkumar8508 11 หลายเดือนก่อน +1

    Thankyou so much for this. God bless you

  • @knutjagersberg381
    @knutjagersberg381 11 หลายเดือนก่อน +1

    Do you really have to get access by Meta to use the weights? My current interpretation is that you enter the license agreement as soon as you use the weights, where ever you got them (as you're also allowed to redistribute them).
    I'm not 100% sure about this, but I think you don't need to register. I think that's more for them to keep track of early adopters.

  • @Nick_With_A_Stick
    @Nick_With_A_Stick 11 หลายเดือนก่อน +2

    My computer is currently training a lora on stable 7b for natural language to (30k)python, and (30k)sql. I also Included (30k)orca questions so it dosent loose its abilities as a language model, and 20k sentiment analysis for new headlines. I would love to try this model with this as soon as Its done training.

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Noiceee, what data sets you using for Python?

  • @randomthoughts7838
    @randomthoughts7838 10 หลายเดือนก่อน

    Hey, is there some structured way(steps) to learn to work with llms. As an analogy, DSA is one structured way to solve coding problems. I am new to llms realm and any advice is much appreciated.

  • @Ryan-yj4sd
    @Ryan-yj4sd 11 หลายเดือนก่อน +1

    Nice video. You seem to have the taken the tough route. I didn't have as much trouble :)

  • @lashlarue59
    @lashlarue59 10 หลายเดือนก่อน

    Nick you said that you were able to build your lip reading model in 96 epochs. How long in an epoch in real time?

  • @strangnet
    @strangnet 9 หลายเดือนก่อน

    Really interesting, but what was your total cost in the end?

  • @Kingupon
    @Kingupon 10 หลายเดือนก่อน

    I am saying Do I need to know metric level math to get Ahead in machine learning or just Know how things work like the specific or library I'm using Pls answer my Question

  • @user-cy4ld4cx1c
    @user-cy4ld4cx1c 4 หลายเดือนก่อน

    I love you nicholas....you are awesome .my only regret is that i didn't found you earlier. all my dream projects in a channel....thankyou

  • @muradahmad9357
    @muradahmad9357 3 หลายเดือนก่อน

    can you please tell which cuda version and nvidia driver versions you used, I am having problem downloading it

  • @pantherg4236
    @pantherg4236 11 หลายเดือนก่อน

    What is the best way to learn deep learning fundamentals via implementation (let's say pick a trivial problem of build a recommendation system for movies) using pytorch in Aug 26, 2023? Thanks in advance

  • @hebjies
    @hebjies 9 หลายเดือนก่อน +4

    It is possible that when you tried to load the pdf with SimpleDirectoryReader, it was skipping pages, because of the chunk size /embedding model you selected, the model you selected (all-MiniLM-L6-v2) is limited to 384 while the chunk you specified was 1024, maybe and just maybe, that is why I think it was skipping pages, because it was unable to load all the chunk in the embedding model

  • @jyothishkumar.j3619
    @jyothishkumar.j3619 10 หลายเดือนก่อน

    What are the limation on monetizing Ilama Banker app ? Could please explain?

  • @dkhundley
    @dkhundley 11 หลายเดือนก่อน +1

    Well done! Have you considered a video around a formal fine tuning of one of the lesser variants (e.g. 7B) version of Llama 2? I’d love to see you do one. 😁

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน +4

      On the cards for this week DK, had a client ask for it. Actually got a super interesting use case in mind!

    • @dkhundley
      @dkhundley 11 หลายเดือนก่อน

      @@NicholasRenotte Awesome! Looking forward to it.

  • @zamirkhurshid261
    @zamirkhurshid261 11 หลายเดือนก่อน +1

    Nice sharing sir your way of teching is very helpful for biggner. Please make a video how we can make deep learning model on earthquake dataset as you have make a project on image classification.

  • @krishnakompalli2606
    @krishnakompalli2606 9 หลายเดือนก่อน

    As you have used RAG method, I'd like to know how it can answer extrapolated questions?

  • @micbab-vg2mu
    @micbab-vg2mu 11 หลายเดือนก่อน +1

    Great video - thank you

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Thanks a mil for checking it ou!

  • @angelazhang9082
    @angelazhang9082 3 หลายเดือนก่อน

    Hi Nick... really late but would be super grateful for a response. I'm trying to figure out how you used RunPOD for this. It looks like you created a folder to store the weights instead of using one of their custom LLM options. Did you pay for extra storage? I can't imaging you loaded all the weights each time you needed to use this on the cloud. I'm new to working with these models and cloud GPUs, so any help is greatly appreciated!

  • @deadcrypt
    @deadcrypt 9 หลายเดือนก่อน

    8:57 nice auth key you got there

  • @Bliss_99988
    @Bliss_99988 11 หลายเดือนก่อน +1

    'How to start a farm with no experience' - Hahaha, man, I just want to say that I love your sense of humour. Also, your videos are really useful for me, I'm an English teacher and I'm trying to build useful tools for my students. Thanks for your content.

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      😂 it's my secret dream job! Hahah thanks so much for checking it out man!!

  • @user-ht9st4up8q
    @user-ht9st4up8q 7 หลายเดือนก่อน

    Gosh when i use GPT4 , it give me a response saying it can not further summarize personal report and it just stop there.
    I think i will just need to switch to a diff models

  • @warthog123
    @warthog123 2 หลายเดือนก่อน

    Excellent video

  • @user-gx2wq5qw8n
    @user-gx2wq5qw8n 10 หลายเดือนก่อน

    Hello nicholas,i still not understand the ./model.

  • @jasonkocher3513
    @jasonkocher3513 11 หลายเดือนก่อน +1

    ...Sunday morning after a bender hahhaaha bro I love you.

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Best time to deploy imho 😅

  • @mrrfrooty
    @mrrfrooty หลายเดือนก่อน

    Hi, could you provide the runpod source code for this? Can't find any outside documentation on how you made this possible

  • @sunkarashreeshreya451
    @sunkarashreeshreya451 11 หลายเดือนก่อน

    You are brilliant. I've been trying to find a tutorial for slidebot.. could you work on it ?

  • @Shishir_Rahman_vg
    @Shishir_Rahman_vg 5 หลายเดือนก่อน

    I have learned enter intermediate level machine learning, now can I start deep learning along with machine learning. please sir tell me

  • @sergeyfedatsenka7201
    @sergeyfedatsenka7201 10 หลายเดือนก่อน

    Does anyone know if renting GPU is cheaper than using Open AI API? By how much? Thank Nicholas for your great content!

  • @evanfreethy8375
    @evanfreethy8375 8 หลายเดือนก่อน

    Wheres the code for the front end website?

  • @leonardoariewibowo7867
    @leonardoariewibowo7867 10 หลายเดือนก่อน

    do you use linux? because i cant run this with my windows machine, bitandbytes didn't support windows for cuda >= 11.0

  • @zakaria20062
    @zakaria20062 9 หลายเดือนก่อน

    Waiting open source like function call ChatGpT will be amazing

  • @ConsultingjoeOnline
    @ConsultingjoeOnline 4 หลายเดือนก่อน

    What version of langchain and langchain-community were you using? Can't get it to work for the life of me.

  • @thepirate_kinz1509
    @thepirate_kinz1509 10 หลายเดือนก่อน

    Can we have a tutorial on conditional GANs please. And multi feature conditional gans as well 😊

  • @fur1ousBlob
    @fur1ousBlob 11 หลายเดือนก่อน +1

    I wanted to use llama in a chatbot. Do you know if that will be possible? I want to know your opinion. I am using rasa framework to build the chatbot but I am not sure how to integrate it.

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Sure can! Seen this? forum.rasa.com/t/how-to-import-huggingface-models-to-rasa/50238

  • @synthclub
    @synthclub 5 หลายเดือนก่อน

    Really cool llama application. Really Impressive.

  • @ahmadshabaz2724
    @ahmadshabaz2724 11 หลายเดือนก่อน +1

    How i get free gpu in web server. I don't have gpu.

  • @youssefghaouipearls
    @youssefghaouipearls 19 วันที่ผ่านมา

    Hello, This seems like a less expensive approach than using Google Cloud. How much did it cost?

  • @scottcurry3767
    @scottcurry3767 11 หลายเดือนก่อน +1

    RunPod A100 instances are looking scarce, any tips on how to adapt for multiple GPU instances?

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Going to give it a crack this week, i've got a fine tuning project coming up. Will let you know. The other option is to use the GGML/4 bit quantized models, reduces the need for such a beefy instance. Also, check out RunPod Secure Cloud, a little pricier but seems to have more availability (I ended up using SC when I was recording results for this vid because the community instances were all unavailable). Not sponsored just in case I'm giving off salesy vibes.

  • @user-vv3jd2qp8w
    @user-vv3jd2qp8w 11 หลายเดือนก่อน +1

    What are the difference between the meta released llama 2 models , hf models and quantised model (ggml) files found in the hugging face? Why cant we use the meta/llama-2-70b model ?

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน +1

      You could! llama-2-70b is the base model, chat is the model fine-tuned for chat. the GGML model is a quantized model (optimized for running on less powerful machines). The hf suffix indicates that it's been updated to run with the transformers library.

    • @poornipoornisha5616
      @poornipoornisha5616 11 หลายเดือนก่อน

      @@NicholasRenotte The 70b chat model downloaded from meta has consolidated.pth files in it. How to use the files to finetune the model for custom datasets ?

  • @cihan1403
    @cihan1403 11 หลายเดือนก่อน +1

    A deep learning in Pytorch video pleaseee

  • @eddysaoudi253
    @eddysaoudi253 11 หลายเดือนก่อน +1

    Hi! Nice video! Is it possible to use llama2 to build an app like autogpt or gpt researcher in a local environment?

    • @Dave-nz5jf
      @Dave-nz5jf 11 หลายเดือนก่อน +2

      Essentially runpod is a local environment. It's a linux server in the cloud , but it's no different than local linux server.

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน +1

      Yup, what he said ^!

  • @farseen1573
    @farseen1573 3 หลายเดือนก่อน

    What platform you are using for 1.69$/hr gpu? Cant find any good gpu cloud providers🥺

  • @horane
    @horane 11 หลายเดือนก่อน

    minute 4:45 comment is confirmation clutch! Never give up!

  • @dgfilms1712
    @dgfilms1712 11 หลายเดือนก่อน +1

    Thank you brother

  • @tejaskumarreddyj3133
    @tejaskumarreddyj3133 2 วันที่ผ่านมา

    Can you please make a video explaining what is the LLM to use when developing a RAG!! It would be of great help if you could make one and also please tell us about how to run this locally on linux!!😁

  • @cihan1403
    @cihan1403 11 หลายเดือนก่อน +1

    You make me love machine learning more

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน +1

      My job here is done 🙌🏼

  • @americanswan
    @americanswan 6 หลายเดือนก่อน

    Can someone explains to me the money required to run an AI application on my local machine?

  • @vikassalaria24
    @vikassalaria24 11 หลายเดือนก่อน

    i am getting error:ValidationError: 1 validation error for HuggingFaceLLM query_wrapper_prompt str type expected (type=type_error.str). I am using 7b chat llama2 model

    • @divyanshumishra6739
      @divyanshumishra6739 10 หลายเดือนก่อน

      Did you resolved that error? I am gettting same error and iam unable to solve it

  • @frazuppi4897
    @frazuppi4897 8 หลายเดือนก่อน

    TL;DR basic RAG with Llama 70B, nothing more, nothing less - (thanks a lot for the video, really well done)

  • @eel789
    @eel789 11 หลายเดือนก่อน +1

    How do I use this with a react frontend?

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน +2

      Could wrap the inference side of the app up into an api with FastAPI then just call out to it using axios!

  • @Tripp111
    @Tripp111 7 หลายเดือนก่อน

    Thank you. ❤️🍕

  • @user-wg3rr9jh9h
    @user-wg3rr9jh9h 6 หลายเดือนก่อน

    You are marvelous! I bow down after witnessing your next level hacking skills 🧐.

  • @nimaheydarzadeh343
    @nimaheydarzadeh343 11 หลายเดือนก่อน +1

    it's great , I try to find like this

  • @nfic5856
    @nfic5856 11 หลายเดือนก่อน +2

    How can it be scalable, since this deployment costs like 2$ per hour? Thanks.

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน +1

      Didn't show it here but if I were scaling this out, the whole thing wouldn't be running on a GPU. The app would be on a lightweight machine and the LLM running on serverless GPU endpoints.

    • @Ryan-yj4sd
      @Ryan-yj4sd 11 หลายเดือนก่อน +1

      @@NicholasRenotte but you would still need to pay to rent an A100 GPU which around $1 to $4 per hour

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Yeah, no real way around that, gotta host somewhere! Especially so if you want to be able to use your own fine-tuned model eventually (coming up soon)!

    • @nfic5856
      @nfic5856 11 หลายเดือนก่อน

      Does gpt3.5-turbo (4k or 16k context) remain cheaper in a small production scale?

  • @malice112
    @malice112 11 หลายเดือนก่อน +1

    I am confused is Llama 2 an LLM or did you use the Huggingface LLM ?

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน +1

      LLaMA 2 70b is the LLM, we loaded it here using the Hugging Face library.

  • @vitalis
    @vitalis 11 หลายเดือนก่อน +1

    can you do a video about analysing trends from websites such as WGSN?

  • @AraShiNoMiwaKo
    @AraShiNoMiwaKo 5 หลายเดือนก่อน

    Any updates?

  • @user-pp4ts5ob1u
    @user-pp4ts5ob1u 11 หลายเดือนก่อน +1

    Excelet video, you are amazing, please update the video "AI Face Body and Hand Pose Detection with Python and Mediapipe", I can't solve the errors, it would be very useful for my university projects, thank you very much.

  • @mfundomonchwe1313
    @mfundomonchwe1313 11 หลายเดือนก่อน +1

    This is awesome!

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน +1

      Thanks a mil!!

    • @mfundomonchwe1313
      @mfundomonchwe1313 11 หลายเดือนก่อน +1

      please attempt the DAG context {model} next, would love to see that, sort of like Causal inference model@@NicholasRenotte

  • @moshekaufman7103
    @moshekaufman7103 4 หลายเดือนก่อน

    Hey Nicholas,
    It's a little disappointing that you haven't actually released the final model yet, even though you mentioned it in the video. While showing the source code is a good start, it's not the same as actually providing the finished product.
    Unfortunately, without the final model itself, it's difficult to take your word for it. To build trust and transparency, it would be much better to provide a download link for the model so people can try it out for themselves. This would be a much more impactful way to share your work and allow others to engage with it.
    I hope you'll reconsider and release the final model soon!

  • @MrDataStorm007
    @MrDataStorm007 11 หลายเดือนก่อน

    your are a god thank you !

  • @singledoutpictures6742
    @singledoutpictures6742 10 หลายเดือนก่อน

    You are a weapon!!!

  • @zimpot1690
    @zimpot1690 11 หลายเดือนก่อน +1

    Can you do a machine learning for trading

    • @NicholasRenotte
      @NicholasRenotte  11 หลายเดือนก่อน

      Yah, it's definitely on my list!