Step-by-step guide on how to setup and run Llama-2 model locally

แชร์
ฝัง
  • เผยแพร่เมื่อ 26 ต.ค. 2024

ความคิดเห็น • 39

  • @abubakeribrahim6473
    @abubakeribrahim6473 9 หลายเดือนก่อน +2

    Thank you for the very nice presentation and explanation! I would like to a video with your wonderful explanation where you can tell us how we can fine-tune the base models to one fitting our specific tasks

    • @ycopie1126
      @ycopie1126  9 หลายเดือนก่อน +1

      Noted, will do it. Thanks!

  • @mj_cta
    @mj_cta 8 หลายเดือนก่อน

    24:39 - Fun part of the video, good luck Yash ! Thanks for the video.

  • @Ravi-b8i
    @Ravi-b8i 28 วันที่ผ่านมา

    thanks bro, really helpfull video

  • @weelianglien687
    @weelianglien687 6 หลายเดือนก่อน +3

    Thank you for this hands-on! Initially I tried it on my laptop which although has an NVDA GEFORCE GTX, it can't run very well. Eventually I have to run it on Colab (T4 GPU), though not with adding the following lines to help with the GPU usage (just sharing) :
    !pip install accelerate
    from accelerate import Accelerator
    accelerator = Accelerator()
    device = accelerator.device

    • @22nd.of.may.
      @22nd.of.may. 3 หลายเดือนก่อน

      my model needs 16.2gb of GPU, which in colab is limited to 15gb, do you have any way to fix that?

  • @vasanthnagkv5654
    @vasanthnagkv5654 4 หลายเดือนก่อน

    Thanks! this was my first AI development video watch.

  • @gaspardtissandier3204
    @gaspardtissandier3204 6 หลายเดือนก่อน +1

    Great video, and it is indeed the right translation to french :)

  • @bhaavamritdhaara
    @bhaavamritdhaara 6 หลายเดือนก่อน

    Very helpful. Thanks a lot for this.

  • @CarolinaHernandez-zt6li
    @CarolinaHernandez-zt6li 7 หลายเดือนก่อน +1

    Do you offer any paid consulting? I’m stuck on an installation error.

  • @thamilarasan4048
    @thamilarasan4048 8 หลายเดือนก่อน +2

    Please share you system specs, specially about GPU you are using

  • @jatindhiman448
    @jatindhiman448 3 หลายเดือนก่อน

    Really great explanation.....
    But i am stuck in a problem of getting space on gpu. If i tried this on google collab ,the free version gets collapsed due to all memory usage. Pls suggest me for this solution or list the name of small models that are under 12gb of space & are used for prompting purpose.

  • @mayowaogundipe375
    @mayowaogundipe375 3 หลายเดือนก่อน

    Thanks for your time... Please may I ask how to download coda toolkit on my laptop to support GPU support. The code for Coda or cpu is not working on my laptop

  • @abhishekfnu7455
    @abhishekfnu7455 8 หลายเดือนก่อน

    Thank you so much for this video.
    Could you please let us know how to connect with SQL database to fetch the information and implement semantic analysis?

  • @pradachan
    @pradachan 5 หลายเดือนก่อน

    i'm new to LLM and i just wanted to know that you need all these access for using llama, but when you'd use ollama you just put "ollama run llama2" in the terminal, so whats the difference? they can access it without any explicit access from meta??

  • @harsh2014
    @harsh2014 7 หลายเดือนก่อน

    Thanks for this discussion !

  • @SpartanDemiGod
    @SpartanDemiGod 8 หลายเดือนก่อน +6

    Can you please tell me your PC specs ?

  • @khaitruong9831
    @khaitruong9831 5 หลายเดือนก่อน

    Great video. Thank you ycopie!

  • @rastapopolous8446
    @rastapopolous8446 8 หลายเดือนก่อน

    nice tutorial but how would you do to wait for the prompt.. so we can enter the prompt like what is capital in Indisa and press enter.. then the model should reply.. how to do it.

  • @niklasweiss2557
    @niklasweiss2557 3 หลายเดือนก่อน

    I currently have the problem that it only says "Loading widget..." when I try to run the code and doesn't display the progress bar. Do you possibly know how to fix this?

  • @LeoSRajan
    @LeoSRajan 8 หลายเดือนก่อน

    Thank you so much for your time!!!

  • @jennilthiyam1261
    @jennilthiyam1261 5 หลายเดือนก่อน

    what will we do if i need interactive mode, like having conversation like we do with chatgpt

  • @sarahharte186
    @sarahharte186 8 หลายเดือนก่อน

    Great vid - Thanks a mil! I am getting KeyError: 'llama' when running the script. I have copied in the model name/path from hugging face directly but its still causing an issue - Do you know what the problem could be?

    • @ycopie1126
      @ycopie1126  8 หลายเดือนก่อน

      Will need more details like which line in your code is causing this

    • @sarahharte186
      @sarahharte186 8 หลายเดือนก่อน +1

      @@ycopie1126 Sorry i think it actually was an issue with the version of transformers I had installed - i reinstalled and now the model seems to be downloading successfully - so all good! Appreciate your reply!

  • @lesstalkeatmore9441
    @lesstalkeatmore9441 5 หลายเดือนก่อน

    how to fine tune with our own data sets, like answer the pdf of our own data sets.

  • @rakeshkumarrout2629
    @rakeshkumarrout2629 3 หลายเดือนก่อน

    can i usethis in vscode?

  • @litttlemooncream5049
    @litttlemooncream5049 5 หลายเดือนก่อน

    really subtle! subscribed

  • @mohammedmujtabaahmed490
    @mohammedmujtabaahmed490 7 หลายเดือนก่อน

    ConnectionError: ('Connection aborted.', ConnectionResetError(10054, 'An existing connection was forcibly closed by the remote host', None, 10054, None))
    bro iam getting this error when running on jupyter notebook.
    please help.

  • @samirait-abbou5954
    @samirait-abbou5954 7 หลายเดือนก่อน

    good job!

  • @fabiotarocasalino257
    @fabiotarocasalino257 7 หลายเดือนก่อน

    good vid

  • @sumandas829
    @sumandas829 7 หลายเดือนก่อน

    hf represents human feedback not hugging face

    • @ycopie1126
      @ycopie1126  7 หลายเดือนก่อน

      You can follow this discussion: github.com/facebookresearch/llama/issues/612
      The model card has small difference which states that it's hugging face format.

    • @sumandas829
      @sumandas829 7 หลายเดือนก่อน

      Extremely sorry for doubting, just thought hf should mean human feedback, again I am wrong, sorry for that, good job

    • @ycopie1126
      @ycopie1126  7 หลายเดือนก่อน

      No worries at all. Happy you put it in comments so it would help other people as well 😄

  • @sachinworld_
    @sachinworld_ 6 หลายเดือนก่อน

    ValueError: You are trying to offload the whole model to the disk. Please use the `disk_offload` function instead.
    i got this error

  • @litttlemooncream5049
    @litttlemooncream5049 5 หลายเดือนก่อน

    love your username lol

  • @MarkSikorski-xg7gh
    @MarkSikorski-xg7gh 8 หลายเดือนก่อน

    Hi when running this code I am getting an error File d:\Magister\llama_hugging_face\venv\lib\site-packages\huggingface_hub\utils\_validators.py:110, in validate_hf_hub_args.._inner_fn(*args, **kwargs)
    109 if arg_name in ["repo_id", "from_id", "to_id"]:
    --> 110 validate_repo_id(arg_value)
    112 elif arg_name == "token" and arg_value is not None:
    ...
    )
    (norm): LlamaRMSNorm()
    )
    (lm_head): Linear(in_features=4096, out_features=32000, bias=False)
    )'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
    Did you encounter this problem?

  • @mohammedmujtabaahmed490
    @mohammedmujtabaahmed490 7 หลายเดือนก่อน

    ConnectionError: ('Connection aborted.', ConnectionResetError(10054, 'An existing connection was forcibly closed by the remote host', None, 10054, None))
    bro iam getting this error when running on jupyter notebook.
    please help.