how to use your model in the lagchain agent? I used this but it says llm value is not a valid dict agent = initialize_agent(tools, model, agent="zero-shot-react-description", verbose=True, handle_parsing_errors=True, max_new_tokens=1000)
Hello there, this is exactly what I was looking for. Could you please give resources or any tutorial where details of those functions are discussed? My teammate gave a Kaggle Notebook with the exact same code and I am continuing to make that a conversational chatbot. But since I am brand new to this, I feel lost now.
great video, sweet and simple. However, how can we control the token max limit, and also, do we have the option of separating our messages into a system message and a user message just like in Openai?
Loved your content buddy ❤. Can we keep this Google Colab instance keep running for free and how can we expose this model as an Rest API to use in hosted projects and that too not locally.
Astonishing content Man 🔥🔥🔥 🚀
Thank you mate !
Nice video, good job!
Thank you! Cheers!
thanks for tutorial!!
What is better quantify with "bitsandbytes" o do it with "cllama" GUFF? What is the difference?
Great video , can you make video on finetuning llm with best method.
That's exactly whats planned Venkatesh. stay tuned..
how to use your model in the lagchain agent? I used this but it says llm value is not a valid dict
agent = initialize_agent(tools,
model,
agent="zero-shot-react-description",
verbose=True,
handle_parsing_errors=True,
max_new_tokens=1000)
thx 😀
hi, is there a simple change that can be made to the code to run inference in 8-bit?
Can you make video how to use open source LLM to query structured databse (sql/pandas) for chat
Sure will try to do one.
Hello there, this is exactly what I was looking for. Could you please give resources or any tutorial where details of those functions are discussed?
My teammate gave a Kaggle Notebook with the exact same code and I am continuing to make that a conversational chatbot. But since I am brand new to this, I feel lost now.
Hi Sir,
Could you tell us the mic setup and how you make your videos with such clear qulaity. Thanks
can i use the minstral7b sharded model as a chatbot, so i can ask it questions about specific data i have for example, a book?
Help full video
great video, sweet and simple. However, how can we control the token max limit, and also, do we have the option of separating our messages into a system message and a user message just like in Openai?
thanks for your tutorial. I have question, how to generate output to 32k ?
Loved your content buddy ❤. Can we keep this Google Colab instance keep running for free and how can we expose this model as an Rest API to use in hosted projects and that too not locally.
Hi, I get my token from huggingface but I dont know where I have to put it in colab
Sir, any advice if I use japanese or chinese language for RAG? Thanks
Can we do this type of qunatization with any model?
yes we can do very much. Checkout my tweet on this
twitter.com/rohanpaul_ai/status/1765688184753820073
colab file not found pls give notebook link
Corrected the link in the description, here it is
github.com/rohan-paul/LLM-FineTuning-Large-Language-Models/blob/main/Mistral-7B-Inferencing.ipynb
how to fine tune this
For finetuning checkout this video
th-cam.com/video/6DGYj1EEWOw/w-d-xo.html&ab_channel=Rohan-Paul-AI