but i guess here's one: finetuning qwen2.5-32b on various ai related github prjs that are written in python, with a focus on agentic workflows, maybe individual finetunes for each agent python lib / framework, such as autogen, langchain / langgraph, ...
Sadly this may be the first time I disagree with that sentiment. Haha. I loved it but was looking for a course link at the end because it was too much too fast ;)
@@OilersFlash Yea bro (actually uncle, cuz I'm 15 and your dp shows...), I realized that the viewer will need some pre knowledge of LLMs and its workings and also some pre basic knowledge of fine-tuning! But yea the video was good actually!
This is exactly how we are tuning our open source LLMs, the use of unsloth+LORA is key. Validation of the learning is an adventure. We did this over the same tech stack mentioned here. Very good video, of course, lots of details behind each step that cannot be explained in a short video. Great use of lighting and your pace is excellent. Looking forward to more vids.
it depends, this is useful, just if you already know how to do it and you just need a refresh on the steps. In other words, you want to watch it just if you do not need it
I was already taken by the clear content but the KEY highlight of simply mentioning Conda shows the high quality of your informations. Subscribed and thanks for the tuto.
Every how-to I've found so far glazes over the training datasets. Like here she goes from here is 100k sql dataset, then here are is how the model expects the prompt, but then doesn't show the format of how the dataset is loaded into the model. What does the sql_context look like, what does the sql_prompt look like, what is the sql, and sql_explaination? Fine tuning is one thing but people also need to know how to build and integrate datasets for training.
Great video straight to the point, but could you please elaborate more on feeding custom dataset to the script? What are the steps? You wrote it right on the script or did you load it from a different file? Or did you upload to hugginface and use the token from there? Im confused
Great tutorial! Thanks. Really like the definitions and explanations rather than just glossing over the code.. would love to see a full tutorial series 😮 I'd like to try to code it myself, but it would be good if you could share the code too.
Thank you for video. I followed your approach for finetuning the model text to SQL. When I work on my own database, its performance is not good. Unable to generate SQL query. I even used agent and examples queries our database. I am requesting suggestion from you. Thank you
Can you train 12B on 24GB or is 12B too big? Another question is if you have multi-turn data (conversations) can you finetune on that? The examples I see are for Q:A pairs.
I'm working on a RAG application. I'm using a pdf file as a text data and I have clean the data as well using NLTK. I already have ollama install in my local system and also llama2 model. I embed the data in to vector form using ollama embedded model mxbai-embed-large and store it on chromabd. but if I give the prompt and get response, it give me the response but not according to my data. and how do I finetune the model? I just need guidance and sequence so I work on the project. It's important to me. I need your guidance. please reply. I'm using windows10. Thank you!
@@ibrahimgamal7603 Could be anything such as you are not using a good doc extractor or your embedding model sucks. Changing chunk and overlap size might help. Don't go for finetuning straight away, do some experimentation and figure out where the problem lies exactly otherwise even finetuning won't work.
if I train a model, will the responses be restricted to just my training data/examples? Or does the model's original training also play a role? For example, say I trained it on model to respond with vehicle specs. I feed it a vehicle name and the model returns the vehicle specs. Say I finetune the model with 50 examples (honda civic, toyota camry, volvo xc90, etc).... however, in reality there are 300,000 possible vehicles. Would finetuning on 50 examples help? Sure, I can put the 50 models in the prompt but then I'd just be wasting tokens.
Thanks! Although I do have a question why you show so much of yourself... I admit it's nice and pleasant to look at, however I would prefer to look at the code or things you talk about. Especially when I watch your videos at work I would definitely like to avoid someone thinking that I watch some silly videos during work while I actually watch work related stuff. Anyway, really interesting videos so thanks again! 😊✌
It's a great how to but you missed a "before trainng" and "after training" examples of how the model responded. Good video anyway. Couldn't get much out of it though
why didn't you directly use hugginface trainer to train, why using unsloth? I want to know what was the benefit of using unsloth over hugginface trainer.
Here it is, she says.. For simplicity.... A pip install command.. In a video.. With no pasted text to copy in the description. That's the opposite of simplicity.
Why are there so many videos that are useless as tutorials and give the impression that they are only about self-promotion? Seriously, anyone who understands this quick run-through here doesn't need any more tutorials, and for the vast majority of the rest, it's probably pretty useless.
Good video, but it’s hard to imagine that you you actually sat there and edited out every breath and pause so that the whole video would sound like one massive run on sentence, only to shave off some 20 or 30 seconds of duration.
Great video straight to the point, but could you please elaborate more on feeding custom dataset to the script? What are the steps? You wrote it right on the script or did you load it from a different file? Or did you upload to hugginface and use the token from there? Im confused
What data would you fine tune your LLM on?
what is your hardware ? are you using intel or amd threadripper ?
that depends on the use case... are you asking for what use cases people are aiming to support?
but i guess here's one: finetuning qwen2.5-32b on various ai related github prjs that are written in python, with a focus on agentic workflows, maybe individual finetunes for each agent python lib / framework, such as autogen, langchain / langgraph, ...
youtube need more chanels alike this. great job
I'm using it to fine-tune a Minecraft bot (Mindcraft from Emergent garden)
No over rating, no over talking, streight forward, love it.
Sadly this may be the first time I disagree with that sentiment. Haha. I loved it but was looking for a course link at the end because it was too much too fast ;)
@@OilersFlash Yea bro (actually uncle, cuz I'm 15 and your dp shows...), I realized that the viewer will need some pre knowledge of LLMs and its workings and also some pre basic knowledge of fine-tuning! But yea the video was good actually!
@@siddhubhai2508 it is good ;)
Knowing such a deep technical subject is one thing, but teaching it well is another! well done young lady.
This is exactly how we are tuning our open source LLMs, the use of unsloth+LORA is key. Validation of the learning is an adventure. We did this over the same tech stack mentioned here. Very good video, of course, lots of details behind each step that cannot be explained in a short video. Great use of lighting and your pace is excellent. Looking forward to more vids.
Thank you straight to the point
I usually have some ptsd when looking at tutorials
it depends, this is useful, just if you already know how to do it and you just need a refresh on the steps. In other words, you want to watch it just if you do not need it
amazing quality of editing, sound, video - beyond the programming side!
A pleasure for the eyes and ears to watch!
Great video for people who know coding and local llm but havent finetuned!
I was already taken by the clear content but the KEY highlight of simply mentioning Conda shows the high quality of your informations. Subscribed and thanks for the tuto.
Easiest subscribe of my life, just wanted something easy and straight to the point
Insanely good video!! Straight to the point and great presentation
Every how-to I've found so far glazes over the training datasets. Like here she goes from here is 100k sql dataset, then here are is how the model expects the prompt, but then doesn't show the format of how the dataset is loaded into the model. What does the sql_context look like, what does the sql_prompt look like, what is the sql, and sql_explaination? Fine tuning is one thing but people also need to know how to build and integrate datasets for training.
Five minutes made useful. Thank you for the crisp and neat video. ❤
This is what I am thinking to do. Finding this video maybe saved half a day
Great video straight to the point, but could you please elaborate more on feeding custom dataset to the script? What are the steps? You wrote it right on the script or did you load it from a different file? Or did you upload to hugginface and use the token from there? Im confused
Great explanation and presentation of LLM.
Hi there.
It's wonderful.. Will you Please share the notebook and also Google Collab notebook?
Appreciate the brevity. Thank you.
good guide to play by myself
wouldve been nice if you had shared the full collab code...
Great tutorial! Thanks. Really like the definitions and explanations rather than just glossing over the code.. would love to see a full tutorial series 😮
I'd like to try to code it myself, but it would be good if you could share the code too.
What a clean presentation.
Intresting video!!
Yoo, what theme are you using in your system?? That's really cool
nice and straight forward approach
I actually liked the music. Great editing also. I found the proyect idea a little bit boring. You could train it for something actually awesome
great and short video guys incredible!!
Well explained. Thanks
Great video. To the point. Effective
Great video! Subscribed! 🎉
Thank you for video. I followed your approach for finetuning the model text to SQL. When I work on my own database, its performance is not good. Unable to generate SQL query. I even used agent and examples queries our database. I am requesting suggestion from you. Thank you
Excellent tutorial! Doesn’t lowering the bit depth of the model greatly reduce accuracy? What are the pros and cons of doing so? Thanks!
Another great video!! Thanks.
Excellent direct video on fine-tuning, congrats. Could you also share the python source code that you used in the video? Thanks.
How long did it take for the training / fine-tuning on your 4090? Thanks for the video!
Isn't this RAG? Doesn't chunking fit anywhere?
what ubuntu are you using, the terminal looks dope
Thanks 🙏
Can you train 12B on 24GB or is 12B too big?
Another question is if you have multi-turn data (conversations) can you finetune on that? The examples I see are for Q:A pairs.
I have 55K classes with an ungodly amount of transcripts. What'll be better and faster? RAG or finetuning?
It was a great video, but I have a question, is it compatible with the new versions called 3.2, especially versions 1B and 3B?
I'm working on a RAG application. I'm using a pdf file as a text data and I have clean the data as well using NLTK. I already have ollama install in my local system and also llama2 model. I
embed the data in to vector form using ollama embedded model mxbai-embed-large and store it on chromabd. but if I give the prompt and get response, it give me the response but not according to my data. and how do I finetune the model? I just need guidance and sequence so I work on the project. It's important to me. I need your guidance. please reply. I'm using windows10.
Thank you!
It's easy you should write in your prompt template = """Answer the question based ONLY on the following context:
{context}
Question: {question}
"""
@@ibrahimgamal7603 Could be anything such as you are not using a good doc extractor or your embedding model sucks. Changing chunk and overlap size might help. Don't go for finetuning straight away, do some experimentation and figure out where the problem lies exactly otherwise even finetuning won't work.
@@muhammadumarnawaz9200 ok mate thank you for your help
fine tuneing offline ?
where is the link for google colab?
Is it possible to make your own AI model with this that speaks to you like Jarvis and understands tone, etc?
if I train a model, will the responses be restricted to just my training data/examples? Or does the model's original training also play a role?
For example, say I trained it on model to respond with vehicle specs. I feed it a vehicle name and the model returns the vehicle specs. Say I finetune the model with 50 examples (honda civic, toyota camry, volvo xc90, etc).... however, in reality there are 300,000 possible vehicles. Would finetuning on 50 examples help? Sure, I can put the 50 models in the prompt but then I'd just be wasting tokens.
what is the estimate cost for training such a model?
can this be done all from the wsl terminal? I am using ollama, webui and docker, but would like to train some company pdf's better, can this be done?
Thanks! Although I do have a question why you show so much of yourself... I admit it's nice and pleasant to look at, however I would prefer to look at the code or things you talk about. Especially when I watch your videos at work I would definitely like to avoid someone thinking that I watch some silly videos during work while I actually watch work related stuff. Anyway, really interesting videos so thanks again! 😊✌
Do you have github repo of the code?
Does fine-tuning shouldn't generate small models that theoretically run fully local and avoid spending money on OpenAI?
It's a great how to but you missed a "before trainng" and "after training" examples of how the model responded. Good video anyway. Couldn't get much out of it though
Apologies if this is a dumb question, but where is, is there a link to the notebook?
Video on formatting data sets to follow? 😅
How to collect dataset pls show it
subscribed
Is unsloath free?
why didn't you directly use hugginface trainer to train, why using unsloth? I want to know what was the benefit of using unsloth over hugginface trainer.
No benefit My friend :(
Give us the collab link! Or better yet the llamasql gguf link!
Unsloth has a dependency of triton, which doesn't seem to be compatible with windows.
Nice
damn, now i want to create my own personal DAN assistant without OpenAI interfere with her
Can we train tinyllama to do something similar? Since was trying to run AI on Raspberry Pi 5 (with Hailo AI Accelerator)
Absolutely!
@@warpdotdev Thanks! ❤️ from India
at 2:00 is 8 Billion not 8 bit.
Nice video, but the music is way too disctracting.
Why not share the source code???
Here it is, she says.. For simplicity.... A pip install command.. In a video.. With no pasted text to copy in the description.
That's the opposite of simplicity.
Congrats for finding such a smartie-cutie as a DevRel for Warp
link
We need a video to explain this video
Can I do this without Conda? I hate Conda
Why yes, why wouldn’t you?
Thankyou sis
She's cute and I can now fine tune my llama 3. Yay 🎉
cant use claude!!!!
let us use ollama with warp ai :(
Ashley, look at me
Why are there so many videos that are useless as tutorials and give the impression that they are only about self-promotion? Seriously, anyone who understands this quick run-through here doesn't need any more tutorials, and for the vast majority of the rest, it's probably pretty useless.
Gonzalez Susan Walker Larry Martinez Joseph
Next time pls avoid the background music , it's distracting from focusing.
Good video, but it’s hard to imagine that you you actually sat there and edited out every breath and pause so that the whole video would sound like one massive run on sentence, only to shave off some 20 or 30 seconds of duration.
Rodriguez John Thomas Lisa Rodriguez Jose
Are you a Mr or a Mrs? It's 202024 I don't wanna assume
20,000 years from now
Windows 🥴
Obama use Ollama
I want to create on that creates computer virus, and hacking software would it be the same idea? lol jk
wow to much stuf in 5min
Please bring it for windowsssssss😢😢😢
shes pretty... what was this video about?
This video is not recommended for the newcomers. She speak fast, in every few second is a cut on the footage. Really not detailed
please talk to the lens and not yourself on the screen you're looking at
White Donald Allen Deborah Anderson Kimberly
You are so pretty.
Why she looks like Andrew Ng 😂😂
uv > anaconda
Build 100% local ChatGPT Clone | Langchain | Part 1 #chatgpt #langchain #langgraph #genai
th-cam.com/video/RAAeWIVvOAQ/w-d-xo.html
Great video straight to the point, but could you please elaborate more on feeding custom dataset to the script? What are the steps? You wrote it right on the script or did you load it from a different file? Or did you upload to hugginface and use the token from there? Im confused