Very interesting i listened to the very end, and it gave me some ideas for prepping my Model. Thanks for the explanation and demo. May I ask?... Do you think an Nvidia GTX Orin Devkit 64 GB would be fitting for running LLMs locally for fine tuning, training and later deploying to server once developed (both Locally and Server on Ubuntu)?
An excellent one! Thank you so much for sharing. Any idea about the possibility of fine tuning my own LLM(like Llama/Mistral), uploading back to HF and the put it into production using VLLM?
Thanks for your video. It is interesting. I am new to LLM and one question to ask. When you run JarvisLabs in your demo, does it mean you are running a server running locally to provide API endpoint? Please advise
In the demo, I was running on a gpu powered instance. The vllm server in this case is running in the Jarvislabs instance. You can use the API endpoint from anywhere.
@@JarvislabsAI thanks so much. I have Navida T500 GPU card on my laptop. But it has only 4 gb. Can it run vLLM? Do we need to install JarvislabsAI on our local machine? Does JarvisLab do? Thanks
hey i also have an AI channel, i tried mistrals model and it didnt finish its execution and looped over the input forever, i had slightly better luck with the instruct version. did you ever get mistral to work?
This was a nicely paced and clear tutorial. Thank you. Liked and subscribed.
Thanks for the support :)
Very interesting i listened to the very end, and it gave me some ideas for prepping my Model. Thanks for the explanation and demo. May I ask?... Do you think an Nvidia GTX Orin Devkit 64 GB would be fitting for running LLMs locally for fine tuning, training and later deploying to server once developed (both Locally and Server on Ubuntu)?
Have not tried it. No idea.
Super useful. Thanks for breaking it down.
Thank you for sharing this information.
Glad it was helpful!
Its a wonderful videa, clearly and concisely explained.
Glad you liked it
An excellent one! Thank you so much for sharing.
Any idea about the possibility of fine tuning my own LLM(like Llama/Mistral), uploading back to HF and the put it into production using VLLM?
Yeah definitely possible. Would make one soon.
@@JarvislabsAI Thank you, looking forward!
hank you, it was interesting.
Thanks for your video. It is interesting.
I am new to LLM and one question to ask.
When you run JarvisLabs in your demo, does it mean you are running a server running locally to provide API endpoint?
Please advise
In the demo, I was running on a gpu powered instance. The vllm server in this case is running in the Jarvislabs instance. You can use the API endpoint from anywhere.
@@JarvislabsAI thanks so much.
I have Navida T500 GPU card on my laptop. But it has only 4 gb. Can it run vLLM?
Do we need to install JarvislabsAI on our local machine?
Does JarvisLab do?
Thanks
Not sure, if will be possible to run vllm on T500 GPU. Jarvislabs, offers a gpu instance in which you can use vllm.
hey i also have an AI channel, i tried mistrals model and it didnt finish its execution and looped over the input forever, i had slightly better luck with the instruct version. did you ever get mistral to work?
We tried with vLLM and remember it working. I will probably check again.