Thank you for the very nice presentation and explanation! I would like to a video with your wonderful explanation where you can tell us how we can fine-tune the base models to one fitting our specific tasks
Thank you for this hands-on! Initially I tried it on my laptop which although has an NVDA GEFORCE GTX, it can't run very well. Eventually I have to run it on Colab (T4 GPU), though not with adding the following lines to help with the GPU usage (just sharing) : !pip install accelerate from accelerate import Accelerator accelerator = Accelerator() device = accelerator.device
Really great explanation..... But i am stuck in a problem of getting space on gpu. If i tried this on google collab ,the free version gets collapsed due to all memory usage. Pls suggest me for this solution or list the name of small models that are under 12gb of space & are used for prompting purpose.
Thanks for your time... Please may I ask how to download coda toolkit on my laptop to support GPU support. The code for Coda or cpu is not working on my laptop
Thank you so much for this video. Could you please let us know how to connect with SQL database to fetch the information and implement semantic analysis?
i'm new to LLM and i just wanted to know that you need all these access for using llama, but when you'd use ollama you just put "ollama run llama2" in the terminal, so whats the difference? they can access it without any explicit access from meta??
nice tutorial but how would you do to wait for the prompt.. so we can enter the prompt like what is capital in Indisa and press enter.. then the model should reply.. how to do it.
I currently have the problem that it only says "Loading widget..." when I try to run the code and doesn't display the progress bar. Do you possibly know how to fix this?
Great vid - Thanks a mil! I am getting KeyError: 'llama' when running the script. I have copied in the model name/path from hugging face directly but its still causing an issue - Do you know what the problem could be?
@@ycopie1126 Sorry i think it actually was an issue with the version of transformers I had installed - i reinstalled and now the model seems to be downloading successfully - so all good! Appreciate your reply!
ConnectionError: ('Connection aborted.', ConnectionResetError(10054, 'An existing connection was forcibly closed by the remote host', None, 10054, None)) bro iam getting this error when running on jupyter notebook. please help.
You can follow this discussion: github.com/facebookresearch/llama/issues/612 The model card has small difference which states that it's hugging face format.
Hi when running this code I am getting an error File d:\Magister\llama_hugging_face\venv\lib\site-packages\huggingface_hub\utils\_validators.py:110, in validate_hf_hub_args.._inner_fn(*args, **kwargs) 109 if arg_name in ["repo_id", "from_id", "to_id"]: --> 110 validate_repo_id(arg_value) 112 elif arg_name == "token" and arg_value is not None: ... ) (norm): LlamaRMSNorm() ) (lm_head): Linear(in_features=4096, out_features=32000, bias=False) )'. Please provide either the path to a local folder or the repo_id of a model on the Hub. Did you encounter this problem?
ConnectionError: ('Connection aborted.', ConnectionResetError(10054, 'An existing connection was forcibly closed by the remote host', None, 10054, None)) bro iam getting this error when running on jupyter notebook. please help.
Thank you for the very nice presentation and explanation! I would like to a video with your wonderful explanation where you can tell us how we can fine-tune the base models to one fitting our specific tasks
Noted, will do it. Thanks!
24:39 - Fun part of the video, good luck Yash ! Thanks for the video.
thanks bro, really helpfull video
Thank you for this hands-on! Initially I tried it on my laptop which although has an NVDA GEFORCE GTX, it can't run very well. Eventually I have to run it on Colab (T4 GPU), though not with adding the following lines to help with the GPU usage (just sharing) :
!pip install accelerate
from accelerate import Accelerator
accelerator = Accelerator()
device = accelerator.device
my model needs 16.2gb of GPU, which in colab is limited to 15gb, do you have any way to fix that?
Thanks! this was my first AI development video watch.
Great video, and it is indeed the right translation to french :)
Very helpful. Thanks a lot for this.
Do you offer any paid consulting? I’m stuck on an installation error.
Please share you system specs, specially about GPU you are using
Really great explanation.....
But i am stuck in a problem of getting space on gpu. If i tried this on google collab ,the free version gets collapsed due to all memory usage. Pls suggest me for this solution or list the name of small models that are under 12gb of space & are used for prompting purpose.
Thanks for your time... Please may I ask how to download coda toolkit on my laptop to support GPU support. The code for Coda or cpu is not working on my laptop
Thank you so much for this video.
Could you please let us know how to connect with SQL database to fetch the information and implement semantic analysis?
i'm new to LLM and i just wanted to know that you need all these access for using llama, but when you'd use ollama you just put "ollama run llama2" in the terminal, so whats the difference? they can access it without any explicit access from meta??
Thanks for this discussion !
Can you please tell me your PC specs ?
Great video. Thank you ycopie!
nice tutorial but how would you do to wait for the prompt.. so we can enter the prompt like what is capital in Indisa and press enter.. then the model should reply.. how to do it.
I currently have the problem that it only says "Loading widget..." when I try to run the code and doesn't display the progress bar. Do you possibly know how to fix this?
Thank you so much for your time!!!
what will we do if i need interactive mode, like having conversation like we do with chatgpt
Great vid - Thanks a mil! I am getting KeyError: 'llama' when running the script. I have copied in the model name/path from hugging face directly but its still causing an issue - Do you know what the problem could be?
Will need more details like which line in your code is causing this
@@ycopie1126 Sorry i think it actually was an issue with the version of transformers I had installed - i reinstalled and now the model seems to be downloading successfully - so all good! Appreciate your reply!
how to fine tune with our own data sets, like answer the pdf of our own data sets.
can i usethis in vscode?
really subtle! subscribed
ConnectionError: ('Connection aborted.', ConnectionResetError(10054, 'An existing connection was forcibly closed by the remote host', None, 10054, None))
bro iam getting this error when running on jupyter notebook.
please help.
good job!
good vid
hf represents human feedback not hugging face
You can follow this discussion: github.com/facebookresearch/llama/issues/612
The model card has small difference which states that it's hugging face format.
Extremely sorry for doubting, just thought hf should mean human feedback, again I am wrong, sorry for that, good job
No worries at all. Happy you put it in comments so it would help other people as well 😄
ValueError: You are trying to offload the whole model to the disk. Please use the `disk_offload` function instead.
i got this error
love your username lol
Hi when running this code I am getting an error File d:\Magister\llama_hugging_face\venv\lib\site-packages\huggingface_hub\utils\_validators.py:110, in validate_hf_hub_args.._inner_fn(*args, **kwargs)
109 if arg_name in ["repo_id", "from_id", "to_id"]:
--> 110 validate_repo_id(arg_value)
112 elif arg_name == "token" and arg_value is not None:
...
)
(norm): LlamaRMSNorm()
)
(lm_head): Linear(in_features=4096, out_features=32000, bias=False)
)'. Please provide either the path to a local folder or the repo_id of a model on the Hub.
Did you encounter this problem?
ConnectionError: ('Connection aborted.', ConnectionResetError(10054, 'An existing connection was forcibly closed by the remote host', None, 10054, None))
bro iam getting this error when running on jupyter notebook.
please help.